![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer hardware & operating systems > Operating systems & graphical user interfaces (GUIs) > General
This book describes a new class of computing devices which are
becoming omnipresent in every day life. They make information
access and processing easily available for everyone from anywhere
at any time. Mobility, wireless connectivity, di- versity, and
ease-of-use are the magic keywords of Pervasive and Ubiquitous
Computing. The book covers these front-end devices as well as their
operating systems and the back-end infrastructure which integrate
these pervasive components into a seamless IT world. A strong
emphasis is placed on the underlying technologies and standards
applied when building up pervasive solutions. These fundamental
topics include commonly used terms such as XML, WAP, UMTS, GPRS,
Bluetooth, Jini, transcoding, and cryptography, to mention just a
few. Voice, Web Application Servers, Portals, Web Services, and
Synchronized and Device Management are new in the second
edition.
This book constitutes the refereed proceedings of the Cryptographers' Track at the RSA Conference 2013, CT-RSA 2013, held in San Francisco, CA, USA, in February/March 2013. The 25 revised full papers presented were carefully reviewed and selected from 89 submissions. The papers are grouped into topical sections covering: side channel attacks, digital signatures, public-key encryption, cryptographic protocols, secure implementation methods, symmetric key primitives, and identity-based encryption.
This book is a result of the Seventh International Conference on Information Sys tems Development-Methods and Tools, Theory and Practice held in Bled, Slovenia, Sep tember 21-23, 1998. The purpose of the conference was to address issues facing academia and industry when specifying, developing, managing, and improving information comput erized systems. During the past few years, many new concepts and approaches emerged in the Information Systems Development (ISD) field. The various theories, methods, and tools available to system developers also bring problems such as choosing the most effec tive approach for a specific task. This conference provides a meeting place for IS re searchers and practitioners from Eastern and Western Europe as well as from other parts of the world. An objective of the conference is not only to share scientific knowledge and in terests but to establish strong professional ties among the participants. The Seventh International Conference on Information Systems Develop ment-ISD'98 continues the concepts of the first Polish-Scandinavian Seminar on Current Trends in Information Systems Development Methodologies held in Gdansk, Poland in 1988. Through the years, the Seminar developed into the International Conference on In formation Systems Development. ISD'99 will be held in Boise, Idaho. The selection of papers was carried out by the International Program Committee. All papers were reviewed in advance by three people. Papers were judged according to their originality, relevance, and presentation quality. All papers were judged only on their own merits, independent of other submissions.
This book is a result of the ISD'99, Eight International Conference on Infonnation Systems Development-Methods and Tools, Theory, and Practice held August 11-13, 1999 in Boise, Idaho, USA. The purpose of this conference was to address the issues facing academia and industry when specifying, developing, managing, and improving infonnation systems. ISD'99 consisted not only of the technical program represented in these Proceedings, but also of plenary sessions on product support and content management systems for the Internet environment, workshop on a new paradigm for successful acquisition of infonnation systems, and a panel discussion on current pedagogical issues in systems analysis and design. The selection of papers for ISD'99 was carried out by the International Program Committee. Papers presented during the conference and printed in this volume have been selected from submissions after fonnal double-blind reviewing process and have been revised by their authors based on the recommendations of reviewers. Papers were judged according to their originality, relevance, and presentation quality. All papers were judged purely on their own merits, independently of other submissions. We would like to thank the authors of papers accepted for ISD'99 who all made gallant efforts to provide us with electronic copies of their manuscripts confonning to common guidelines. We thank them for thoughtfully responding to reviewers comments and carefully preparing their final contributions. We thank Daryl Jones, provost of Boise State University and William Lathen, dean, College of Business and Economics, for their support and encouragement.
Cooperating Heterogeneous Systems provides an in-depth introduction to the issues and techniques surrounding the integration and control of diverse and independent software components. Organizations increasingly rely upon diverse computer systems to perform a variety of knowledge-based tasks. This presents technical issues of interoperability and integration, as well as philosophical issues of how cooperation and interaction between computational entities is to be realized. Cooperating systems are systems that work together towards a common end. The concepts of cooperation must be realized in technically sound system architectures, having a uniform meta-layer between knowledge sources and the rest of the system. The layer consists of a family of interpreters, one for each knowledge source, and meta-knowledge. A system architecture to integrate and control diverse knowledge sources is presented. The architecture is based on the meta-level properties of the logic programming language Prolog. An implementation of the architecture is described, a Framework for Logic Programming Systems with Distributed Execution (FLiPSiDE). Knowledge-based systems play an important role in any up-to-date arsenal of decision support tools. The tremendous growth of computer communications infrastructure has made distributed computing a viable option, and often a necessity in geographically distributed organizations. It has become clear that to take knowledge-based systems to their next useful level, it is necessary to get independent knowledge-based systems to work together, much as we put together ad hoc work groups in our organizations to tackle complex problems. The book is for scientists and software engineers who have experience in knowledge-based systems and/or logic programming and seek a hands-on introduction to cooperating systems. Researchers investigating autonomous agents, distributed computation, and cooperating systems will find fresh ideas and new perspectives on well-established approaches to control, organization, and cooperation.
An operating system is probably the most important part of the body of soft ware which goes with any modern computer system. I ts importance is reflected in the large amount of manpower usually invested in its construction, and in the mystique by which it is often surrounded. To the non-expert the design and construction of operating systems has often appeared an activity impenetrable to those who do not practise it. I hope this book will go some way toward dispelling the mystique, and encourage a greater general understanding of the principles on which operating systems are constructed. The material in the book is based on a course of lectures I have given for the past few years to undergraduate students of computer science. The book is therefore a suitable introduction to operating systems for students who have a basic grounding in computer science, or for people who have worked with computers for some time. Ideally the reader should have a knowledge of prorramming and be familiar with general machine architecture, common data structures such as lists and trees, and the functions of system software such as compilers, loaders, and editors. I t will also be helpful if he has had some experience of using a large operating system, seeing it, as it were, from the out side."
This book constitutes the refereed post-proceedings of the 9th IFIP International Conference on Network and Parallel Computing, NPC 2012, held in Gwangju, Korea, in September 2012. The 38 papers presented were carefully reviewed and selected from 136 submissions. The papers are organized in the following topical sections: algorithms, scheduling, analysis, and data mining; network architecture and protocol design; network security; paralel, distributed, and virtualization techniques; performance modeling, prediction, and tuning; resource management; ubiquitous communications and networks; and web, communication, and cloud computing. In addition, a total of 37 papers selected from five satellite workshops (ATIMCN, ATSME, Cloud&Grid, DATICS, and UMAS 2012) are included.
Vorwort In der Natur entwickelten sich die Echtzeitsysteme seit einigen 100 Mil- Honen Jahren. Tierische Nervensysteme haben zur Aufgabe, auf die Nachrichten aus der Umwelt die Steuerungsbefehle an die aktiven Or- gane zu geben. Dabei spielen zum Beispiel bedingte Reflexe eine wichtige Rolle. Vielleicht kann man die Entstehung des Menschen etwa zu der Zeit ansetzen, als sein sich allmahlich entwickelndes Gehirn Gedanken entwickelte, deren Bedeutung in vorausplanender Weise iiber die gerade vorliegende Situation hinausging. Das fiihrte schliesslich unter anderem zum heutigen Wissenschaftler, der seine Theorien und Systeme aufgrund langwieriger Uberlegungen aufbaut. Die Entwicklung der Computer ging im wesentlichen den umgekehrten Weg. Zunachst diente sie nur der Durchfiihrung "starrer" Programme, wie z.B. das erste programmgesteuerte Rechengerat Z3, das der Unterzeichner im Jahre 1941 vorfiihren konnte. Es folgte unter an- derem ein Spezialgerat zur Fliigelvermessung, das man als den ersten Prozessrechner bezeichnen kann. Es wurden etwa vierzig als Analog- Digital-Wandler arbeitende Messuhren yom Rechnerautomaten abgele- sen und im Rahmen eines Programms als Variable verarbeitet. Abel' auch das erfolgte noch in starrer Reihenfolge. Die echte Prozesssteuerung - heute auch Echtzeitsysteme genannt - erfordert aber ein Reagieren auf bestandig wechselnde Situationen.
Compiler technology is fundamental to computer science since it provides the means to implement many other tools. It is interesting that, in fact, many tools have a compiler framework - they accept input in a particular format, perform some processing and present output in another format. Such tools support the abstraction process and are crucial to productive systems development. The focus of Compiler Technology: Tools, Translators and Language Implementation is to enable quick development of analysis tools. Both lexical scanner and parser generator tools are provided as supplements to this book, since a hands-on approach to experimentation with a toy implementation aids in understanding abstract topics such as parse-trees and parse conflicts. Furthermore, it is through hands-on exercises that one discovers the particular intricacies of language implementation. Compiler Technology: Tools, Translators and Language Implementation is suitable as a textbook for an undergraduate or graduate level course on compiler technology, and as a reference for researchers and practitioners interested in compilers and language implementation.
Multiple processor systems are an important class of parallel systems. Over the years, several architectures have been proposed to build such systems to satisfy the requirements of high performance computing. These architectures span a wide variety of system types. At the low end of the spectrum, we can build a small, shared-memory parallel system with tens of processors. These systems typically use a bus to interconnect the processors and memory. Such systems, for example, are becoming commonplace in high-performance graph ics workstations. These systems are called uniform memory access (UMA) multiprocessors because they provide uniform access of memory to all pro cessors. These systems provide a single address space, which is preferred by programmers. This architecture, however, cannot be extended even to medium systems with hundreds of processors due to bus bandwidth limitations. To scale systems to medium range i. e. , to hundreds of processors, non-bus interconnection networks have been proposed. These systems, for example, use a multistage dynamic interconnection network. Such systems also provide global, shared memory like the UMA systems. However, they introduce local and remote memories, which lead to non-uniform memory access (NUMA) architecture. Distributed-memory architecture is used for systems with thousands of pro cessors. These systems differ from the shared-memory architectures in that there is no globally accessible shared memory. Instead, they use message pass ing to facilitate communication among the processors. As a result, they do not provide single address space.
Language, Compilers and Run-time Systems for Scalable Computers contains 20 articles based on presentations given at the third workshop of the same title, and 13 extended abstracts from the poster session. Starting with new developments in classical problems of parallel compiler design, such as dependence analysis and an exploration of loop parallelism, the book goes on to address the issues of compiler strategy for specific architectures and programming environments. Several chapters investigate support for multi-threading, object orientation, irregular computation, locality enhancement, and communication optimization. Issues of the interface between language and operating system support are also discussed. Finally, the load balance issues are discussed in different contexts, including sparse matrix computation and iteratively balanced adaptive solvers for partial differential equations. Some additional topics are also discussed in the extended abstracts. Each chapter provides a bibliography of relevant papers and the book can thus be used as a reference to the most up-to-date research in parallel software engineering.
The papers in this volume were presented at the Second Annual Work shop on Active Middleware Services and were selected for inclusion here by the Editors. The AMS workshop was organized with support from both the National Science Foundation and the CAT center at the Uni versity of Arizona, and was held in Pittsburgh, Pennsylvania, on August 1, 2000, in conjunction with the 9th IEEE International Symposium on High Performance Distributed Computing (HPDC-9). The explosive growth of Internet-based applications and the prolifer ation of networking technologies has been transforming most areas of computer science and engineering as well as computational science and commercial application areas. This opens an outstanding opportunity to explore new, Internet-oriented software technologies that will open new research and application opportunities not only for the multimedia and commercial world, but also for the scientific and high-performance computing applications community. Two emerging technologies - agents and active networks - allow increased programmability to enable bring ing new services to Internet based applications. The AMS workshop presented research results and working papers in the areas of active net works, mobile and intelligent agents, software tools for high performance distributed computing, network operating systems, and application pro gramming models and environments. The success of an endeavor such as this depends on the contributions of many individuals. We would like to thank Dr. Frederica Darema and the NSF for sponsoring the workshop.
This book constitutes the refereed proceedings of the 6th International Conference on Mathematical Methods, Models, and Architectures for Computer Network Security, MMM-ACNS 2012, held in St. Petersburg, Russia in October 2012. The 14 revised full papers and 8 revised short presentations were carefully reviewed and selected from a total of 44 submissions. The papers are organized in topical sections on applied cryptography and security protocols, access control and information protection, security policies, security event and information management, instrusion prevention, detection and response, anti-malware techniques, security modeling and cloud security.
In brief summary, the following results were presented in this work: * A linear time approach was developed to find register requirements for any specified CS schedule or filled MRT. * An algorithm was developed for finding register requirements for any kernel that has a dependence graph that is acyclic and has no data reuse on machines with depth independent instruction templates. * We presented an efficient method of estimating register requirements as a function of pipeline depth. * We developed a technique for efficiently finding bounds on register require ments as a function of pipeline depth. * Presented experimental data to verify these new techniques. * discussed some interesting design points for register file size on a number of different architectures. REFERENCES [1] Robert P. Colwell, Robert P. Nix, John J O'Donnell, David B Papworth, and Paul K. Rodman. A VLIW Architecture for a Trace Scheduling Com piler. In Architectural Support for Programming Languages and Operating Systems, pages 180-192, 1982. [2] C. Eisenbeis, W. Jalby, and A. Lichnewsky. Compile-Time Optimization of Memory and Register Usage on the Cray-2. In Proceedings of the Second Workshop on Languages and Compilers, Urbana l/inois, August 1989. [3] C. Eisenbeis, William Jalby, and Alain Lichnewsky. Squeezing More CPU Performance Out of a Cray-2 by Vector Block Scheduling. In Proceedings of Supercomputing '88, pages 237-246, 1988. [4] Michael J. Flynn. Very High-Speed Computing Systems. Proceedings of the IEEE, 54:1901-1909, December 1966.
Organized Human Activity and Its Support by Computer proposes an answer to the question: what are computers for? With technical expertise, Anatol Holt analyzes human activity and its relevance to computer use. Holt interleaves a theory about the universal aspect of social life with a vision of how to harness computer power. This book is a culmination of a life of work that exemplifies two characteristics of the author: intellectual passion, and a concern for what matters to people. In the past thirty years, Holt has been a participant in the computing work at every level, from managing computer systems to developing commercial software to publishing theoretical articles in academic journals. His breadth of knowledge and experience makes possible the interweaving of theory and practice that shapes the fabric of this book. People often make a false opposition between theory and practice. In this case, it is a synergy: practice guides the theory, and the theory is grounded in its application.' Terry Winograd, Stanford University Organized Human Activity and Its Support by Computer will be of interest to those concerned with computers, especially those with and interest in groupware'. Particular relevance to social scientists, management scientists, students of law, and philosophers are also addressed. Though technical in spirit and method, this book does not expect significant prior computer knowledge of the reader.
This book constitutes the thoroughly refereed conference proceedings of the 5th International Conference on Trusted Systems, INTRUST 2013, held in Graz, Austria, in December 2013. The revised full papers focus on the theory, technologies and applications of trusted systems. They cover all aspects of trusted computing systems, including trusted modules, platforms, networks, services and applications, from their fundamental features and functionalities to design principles, architecture and implementation technologies.
Using Windows 8.1 is the essential step by step guide to using a computer running Windows 8, providing a complete resource for both the beginner and the enthusiast. Techniques are illustrated in step-by-step using photography and screen prints throughout, together with concise, easy to follow text from an established expert in the field, provide a comprehensive guide to Windows. Whether you have just bought your first computer, laptop or are a keen computer user who has just upgraded to Windows 8, this book will provide you with a firm grasp of the underpinning foundations and equip you with the skills needed to use a computer like a pro.
This book constitutes the refereed proceedings of the Fourth International Symposium on Search-Based Software Engineering, SSBSE 2012, held in Riva del Garda, Italy in collocation with the 28th IEEE International Conference on Software Maintenance. The 15 revised full papers, 3 revised short papers, and 2 papers of the graduate track presented together with 2 keynote talks and 1 tutorial paper were carefully reviewed and selected from 38 initial submissions. Search-based Software Engineering (SBSE) studies the application of meta-heuristic optimization techniques to various software engineering problems, ranging from requirements engineering to software testing and maintenance. The papers present current research in all areas of Search Based Software Engineering, including theoretical work, research on SBSE applications, empirical studies, and reports on industrial experience.
This book constitutes the thoroughly refereed proceedings of the 18th International Conference, Euro-Par 2012, held in Rhodes Islands, Greece, in August 2012. The 75 revised full papers presented were carefully reviewed and selected from 228 submissions. The papers are organized in topical sections on support tools and environments; performance prediction and evaluation; scheduling and load balancing; high-performance architectures and compilers; parallel and distributed data management; grid, cluster and cloud computing; peer to peer computing; distributed systems and algorithms; parallel and distributed programming; parallel numerical algorithms; multicore and manycore programming; theory and algorithms for parallel computation; high performance network and communication; mobile and ubiquitous computing; high performance and scientific applications; GPU and accelerators computing.
This book constitutes the refereed proceedings of the 6th International Conference on Provable Security, ProvSec 2012, held in Chengdu, China, in September 2012. The 16 full papers and 4 short papers presented were carefully reviewed and selected from 66 submissions. The papers are grouped in topical sections on signature schemes, foundations, leakage resilence and key escrow, encryption schemes, and information theoretical security.
This book constitutes the proceedings of the 15th International Symposium on Research in Attacks, Intrusions and Defenses, former Recent Advances in Intrusion Detection, RAID 2012, held in Amsterdam, The Netherlands in September 2012. The 18 full and 12 poster papers presented were carefully reviewed and selected from 84 submissions. The papers address all current topics in virtualization, attacks and defenses, host and network security, fraud detection and underground economy, web security, intrusion detection.
Ada 2012 is the latest version of the international standard for the programming language Ada. It is designated ISO/IEC 8652:2012 (E) and is a new edition replacing the 2005 version. The primary goals for the new version were to further enhance its capabilities particularly in those areas where its reliability and predictability are of great value. Many important new features have been included such as those defining dynamic contracts and for handling multiprocessors and are integrated within the existing language framework in an elegant and coherent manner. The Ada 2012 Rationale describes not only the changes from Ada 2005 but also the reason for the changes. It starts with an introduction providing a general overview and this is followed by seven chapters focusing on contracts and aspects; extended expressions; structure and visibility; tasking and real time; iterators and pools; predefined library and containers. The book concludes with an epilogue largely concerned with compatibility issues.
I felt deeply honored when Professor Sumit Ghosh asked me to write the foreword to his book with an extraordinary perspective. I have long admired him, ?rst as a student leader at Stanford, where he initiated the ?rst IEEE Computer Society's student chapter, and later as an esteemed and inspiring friend whose transdisciplinary research broadened and enhanced the horizons of practitioners of computer science and engineering, including my own. His ideas, which are derived from his profound vision, deep critical thinking, and personal intuition, reach from information technology to bioscience, as - hibited in this excellent book. To me, an ordinary engineer, it opens up a panoramic view of the Universe of Knowledge that keeps expanding and - spiring,likethegoodIndianproverb,whichsays,"agoodbookinformsyou,an excellent book teaches you, and a great book changes you. " I sincerely believe that Professor Ghosh's book will help us change and advance the methods of systems engineering and technology. Vision Inspired vision sees ahead of others what will or may come to be, a vivid, imagined concept or anticipation. An inspired vision personi?es what is good and what like-minded individuals hope for. Our vision is one of creating an Internet of minds, where minds are Web sites or knowledge centers, which create, store, and radiate knowledge through interaction with other minds connected by a universal shared network. This vision will not just hasten the death of distance, but will also - carcerate ignorance.
This volume presents the refereed proceedings of the 7th International ICST Conference on Security and Privacy in Communication Networks, SecureComm 2011, held in London, UK, in September 2011. The 35 revised papers included in the volume were carefully reviewed and selected from numerous submissions. Topics covered include network intrusion detection; anonymity and privacy; wireless security; system security; DNS and routing security; and key management.
This book introduces the fundamental concepts and practical simulation te- niques for modeling different aspects of operating systems to study their g- eral behavior and their performance. The approaches applied are obje- oriented modeling and process interaction approach to discrete-event simu- tion. The book depends on the basic modeling concepts and is more specialized than my previous book: Practical Process Simulation with Object-Oriented Techniques and C++, published by Artech House, Boston 1999. For a more detailed description see the Web location: http: //science.kennesaw.edu/ jgarrido/mybook, html. Most other books on performance modeling use only analytical approaches, and very few apply these concepts to the study of operating systems. Thus, the unique feature of the book is that it concentrates on design aspects of operating systems using practical simulation techniques. In addition, the book illustrates the dynamic behavior of different aspects of operating systems using the various simulation models, with a general hands-on approac |
You may like...
|