![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer hardware & operating systems > Operating systems & graphical user interfaces (GUIs) > General
Parallel and distributed computing is one of the foremost
technologies for shaping New Horizons of Parallel and Distributed Computing is a collection of self-contained chapters written by pioneering researchers to provide solutions for newly emerging problems in this field. This volume will not only provide novel ideas, work in progress and state-of-the-art techniques in the field, but will also stimulate future research activities in the area of parallel and distributed computing with applications. New Horizons of Parallel and Distributed Computing is intended for industry researchers and developers, as well as for academic researchers and advanced-level students in computer science and electrical engineering. A valuable reference work, it is also suitable as a textbook.
DAPSYS (International Conference on Distributed and Parallel Systems) is an international biannual conference series dedicated to all aspects of distributed and parallel computing. DAPSYS 2008, the 7th International Conference on Distributed and Parallel Systems was held in September 2008 in Hungary. Distributed and Parallel Systems: Desktop Grid Computing, based on DAPSYS 2008, presents original research, novel concepts and methods, and outstanding results. Contributors investigate parallel and distributed techniques, algorithms, models and applications; present innovative software tools, environments and middleware; focus on various aspects of grid computing; and introduce novel methods for development, deployment, testing and evaluation. This volume features a special focus on desktop grid computing as well. Designed for a professional audience composed of practitioners and researchers in industry, this book is also suitable for advanced-level students in computer science.
This book constitutes the refereed proceedings of the 15th International Conference on Principles of Distributed Systems, OPODIS 2011, held in Toulouse, France, in December 2011. The 26 revised papers presented in this volume were carefully reviewed and selected from 96 submissions. They represent the current state of the art of the research in the field of the design, analysis and development of distributed and real-time systems.
This essential resource for professionals and advanced students in security programming and system design introduces the foundations of programming systems security and the theory behind access control models, and addresses emerging access control mechanisms.
This book studies algorithmic issues associated with cooperative execution of multiple independent tasks by distributed computing agents including partitionable networks. It provides the most significant algorithmic solution developed and available today for do-all computing for distributed systems (including partitionable networks), and is the first monograph that deals with do-all computing for distributed systems. The book is structured to meet the needs of a professional audience composed of researchers and practitioners in industry. This volume is also suitable for graduate-level students in computer science.
"The Testing Network" presents an integrated approach to testing based on cutting-edge methodologies, processes and tools in today's IT context. It means complex network-centric applications to be tested in heterogeneous IT infrastructures and in multiple test environments (also geographically distributed). The added-value of this book is the in-depth explanation of all processes and relevant methodologies and tools to address this complexity. Main aspects of testing are explained using TD/QC - the world-leader test platform. This up-to-date know-how is based on real-life IT experiences gained in large-scale projects of companies operating worldwide. The book is abundantly illustrated to better show all technical aspects of modern testing in a national and international context. The author has a deep expertise by designing and giving testing training in large companies using the above-mentioned tools and processes. "The Testing Network" is a unique synthesis of core test topics applied in real-life.
This book constitutes the thoroughly refereed post-conference proceedings of the Second International ICST Conference on Sensor Systems and Software, S-Cube 2010, held in Miami, Florida, USA, in December 2010. The 17 revised full papers presented were carefully reviewed and selected and cover a wide range of topics including sensor application programming paradigms, novel sensor applications, sensor network middleware, trust security and privacy, wireless sensor network management and monitoring, and sensor application development support systems.
This book provides an in-depth description of event-based systems, covering topics ranging from local event matching and distributed event forwarding algorithms, through a practical discussion of software engineering issues raised by the event-based style, to state-of-the-art research in event-based systems like composite event detection and security. The authors offer a comprehensive overview, and show the power of event-based architectures in modern system design, encouraging professionals to exploit this technique in next generation large-scale distributed applications like information dissemination, network monitoring, enterprise application integration, or mobile systems.
This book constitutes the thoroughly refereed post-conference proceedings of the Third International ICST Conference on Forensic Applications and Techniques in Telecommunications, Information and Multimedia, E-Forensics 2010, held in Shanghai, China, in November 2010. The 32 revised full papers presented were carefully reviewed and selected from 42 submissions in total. These, along with 5 papers from a collocated workshop of E-Forensics Law, cover a wide range of topics including digital evidence handling, data carving, records tracing, device forensics, data tamper identification, and mobile device locating.
Intelligent Integration of Information presents a collection of chapters bringing the science of intelligent integration forward. The focus on integration defines tasks that increase the value of information when information from multiple sources is accessed, related, and combined. This contributed volume has also been published as a special double issue of the Journal of Intelligent Information Systems (JIIS), Volume 6:2/3.
Over the last decade, we have witnessed a growing dependency on information technologyresultingina wide rangeofnew opportunities. Clearly, ithas become almost impossible to imagine life without a personal computer or laptop, or without a cell phone. Social network sites (SNS) are competing with face-- face encounters and may even oust them. Most SNS-adepts have hundreds of "friends," happily sharing pictures and pro?les and endless chitchat. We are on the threshold of the Internet of Things, where every object will have its RFID-tag. This will not only e?ect companies, who will be able to optimize their production and delivery processes, but also end users, who will be able to enjoy many new applications, ranging from smart shopping, and smart fridges to geo-localized services. In the near future, elderly people will be able to stay longer at home due to clever health monitoring systems. The sky seems to be the limit However, we have also seen the other side of the coin: viruses, Trojan horses, breaches of privacy, identity theft, and other security threats. Our real and virtual worlds are becoming increasingly vulnerable to attack. In order to encouragesecurity researchby both academia and industry and to stimulate the dissemination of results, conferences need to be organized. With the 11th edition of the joint IFIP TC-6 TC-11 Conference on C- munications and Multimedia Security (CMS 2010), the organizers resumed the tradition of previous CMS conferences after a three-year recess.
As almost no other technology, embedded systems is an essential element of many innovations in automotive engineering. New functions and improvements of already existing functions, as well as the compliance with traffic regulations and customer requirements, have only become possible by the increasing use of electronic systems, especially in the fields of driving, safety, reliability, and functionality. Along with the functionalities that increase in number and have to cooperate, the complexity of the entire system will increase. Synergy effects resulting from distributed application functionalities via several electronic control devies, exchanging information through the network brings about more complex system architectures with many different sub-networks, operating with different velocities and different protocol implementations. To manage the increasing complexity of these systems, a deterministic behaviour of the control units and the communication network must be provided for, in particular when dealing with a distributed functionality. From Specification to Embedded Systems Application documents recent approaches and results presented at the International Embedded Systems Symposium (IESS 2005), which was held in August 2005 in Manaus (Brazil) and sponsored by the International Federation for Information Processing (IFIP). The topics which have been chosen for this working conference
are very timely: design methodology, modeling, specification,
software synthesis, power management, formal verification, testing,
network, communication systems, distributed control systems,
resource management and special aspects in system design.
This book intends to inculcate the innovative ideas for the scheduling aspect in distributed computing systems. Although the models in this book have been designed for distributed systems, the same information is applicable for any type of system. The book will dramatically improve the design and management of the processes for industry professionals. It deals exclusively with the scheduling aspect, which finds little space in other distributed operating system books. Structured for a professional audience composed of researchers and practitioners in industry, this book is also suitable as a reference for graduate-level students.
As the telecommunication industry introduces new sophisticated technologies, the nature of services and the volume of demands have changed. Indeed, a broad range of new services for users appear, combining voice, data, graphics, video, etc. This implies new planning issues. Fiber transmission systems that can carry large amounts of data on a few strands of wire were introduced. These systems have such a large bandwidth that the failure of even a single transmission link: in the network can create a severe service loss to customers. Therefore, a very high level of service reliability is becoming imperative for both system users and service providers. Since equipment failures and accidents cannot be avoided entirely, networks have to be designed so as to "survive" failures. This is done by judiciously installing spare capacity over the network so that all traffic interrupted by a failure may be diverted around that failure by way of this spare or reserve capacity. This of course translates into huge investments for network operators. Designing such survivable networks while minimizing spare capacity costs is, not surprisingly, a major concern of operating companies which gives rise to very difficult combinatorial problems. In order to make telecommunication networks survivable, one can essentially use two different strategies: protection or restoration. The protection approach preas signs spare capacity to protect each element of the network independently, while the restoration approach spreads the redundant capacity over the whole network and uses it as required in order to restore the disrupted traffic."
The control and data flow of a program can be represented using continuations, a concept from denotational semantics that has practical application in real compilers. This book shows how continuation-passing style is used as an intermediate representation on which to perform optimisations and program transformations. Continuations can be used to compile most programming languages. The method is illustrated in a compiler for the programming language Standard ML. However, prior knowledge of ML is not necessary, as the author carefully explains each concept as it arises. This is the first book to show how concepts from the theory of programming languages can be applied to the producton of practical optimising compilers for modern languages like ML. This book will be essential reading for compiler writers in both industry and academe, as well as for students and researchers in programming language theory.
The Extreme Programming Pocket Guide covers XP assumptions, principles, events, artifacts, roles, and resources, and more. It concisely explains the relationships between the XP practices. If you want to adopt XP in stages, the Extreme Programming Pocket Guide will help you choose what to apply and when. Concise and easy to use, this handy pocket guide to XP is a must-have quick reference for anyone implementing a test-driven development environment.
A number of widely used contemporary processors have instruction-set extensions for improved performance in multi-media applications. The aim is to allow operations to proceed on multiple pixels each clock cycle. Such instruction-sets have been incorporated both in specialist DSPchips such as the Texas C62xx (Texas Instruments, 1998) and in general purpose CPU chips like the Intel IA32 (Intel, 2000) or the AMD K6 (Advanced Micro Devices, 1999). These instruction-set extensions are typically based on the Single Instruc tion-stream Multiple Data-stream (SIMD) model in which a single instruction causes the same mathematical operation to be carried out on several operands, or pairs of operands, at the same time. The level or parallelism supported ranges from two floating point operations, at a time on the AMD K6 architecture to 16 byte operations at a time on the Intel P4 architecture. Whereas processor architectures are moving towards greater levels of parallelism, the most widely used programming languages such as C, Java and Delphi are structured around a model of computation in which operations takeplace on a single value at a time. This was appropriate when processors worked this way, but has become an impediment to programmers seeking to make use of the performance offered by multi-media instruction -sets. The introduction of SIMD instruction sets (Peleg et al."
This book constitutes the thoroughly refereed joint post proceedings of two international workshops, the 5th International Workshop on Data Privacy Management, DPM 2010, and the 3rd International Workshop on Autonomous and Spontaneous Security, SETOP 2010, collocated with the ESORICS 2010 symposium in Athens, Greece, in September 2010. The 9 revised full papers for DPM 2010 presented together with two keynote talks are accompanied by 7 revised full papers of SETOP 2010; all papers were carefully reviewed and selected for inclusion in the book. The DPM 2010 papers cover topics such as how to translate the high-level business goals into system-level privacy policies, administration of privacy-sensitive data, privacy data integration and engineering, privacy access control mechanisms, information-oriented security, and query execution on privacy-sensitive data for partial answers. The SETOP 2010 papers address several specific aspects of the previously cited topics, as for instance the autonomic administration of security policies, secure P2P storage, RFID authentication, anonymity in reputation systems, etc.
This book constitutes the refereed proceedings of the Third
International Symposium on NASA Formal Methods, NFM 2011, held in
Pasadena, CA, USA, in April 2011.
As future generation information technology (FGIT) becomes specialized and fr- mented, it is easy to lose sight that many topics in FGIT have common threads and, because of this, advances in one discipline may be transmitted to others. Presentation of recent results obtained in different disciplines encourages this interchange for the advancement of FGIT as a whole. Of particular interest are hybrid solutions that c- bine ideas taken from multiple disciplines in order to achieve something more signi- cant than the sum of the individual parts. Through such hybrid philosophy, a new principle can be discovered, which has the propensity to propagate throughout mul- faceted disciplines. FGIT 2009 was the first mega-conference that attempted to follow the above idea of hybridization in FGIT in a form of multiple events related to particular disciplines of IT, conducted by separate scientific committees, but coordinated in order to expose the most important contributions. It included the following international conferences: Advanced Software Engineering and Its Applications (ASEA), Bio-Science and Bio-Technology (BSBT), Control and Automation (CA), Database Theory and Application (DTA), D- aster Recovery and Business Continuity (DRBC; published independently), Future G- eration Communication and Networking (FGCN) that was combined with Advanced Communication and Networking (ACN), Grid and Distributed Computing (GDC), M- timedia, Computer Graphics and Broadcasting (MulGraB), Security Technology (SecTech), Signal Processing, Image Processing and Pattern Recognition (SIP), and- and e-Service, Science and Technology (UNESST).
Explains fault tolerance in clear terms, with concrete examples
drawn from real-world settings
At the School of Information Technology, KMUTT, we believe that information te- nology is the most important driver of economy and social development. IT can - able better productivity, as well as helping us to save resources. IT is giving rise to a new round of industrial and business revolution. We now can have products and s- vices that once were believed to be beyond reach. Without IT, it is impossible for people to realize their full potential. Businesses worldwide are harnessing the power of broadband communication, which will have a profound and constructive impact on the economic, social devel- ment, education, and almost all aspects of our life. This new era of unified commu- cation presents us with new challenges. This is why we should work together more closely to enhance the exchange of knowledge related to effective application of broadband communication and IT. It is my sincere hope that all contributions to the Third International Conference on Advances in Information Technology (IAIT 2009) will increase our understanding of how we can have effectively apply this emerging technology for the benefit of all people all around the world. I hope IAIT 2009 will also lead to more research that can contr- ute to a better methodology for IT applications in the era of unified communication. I am very grateful to all our keynotes speakers for coming all the way to Thailand.
This edition marks the tenth Middleware conference. The ?rst conference was held in the Lake District of England in 1998, and its genesis re?ected a growing realization that middleware systems were a unique breed of distributed system requiring their own rigorous research and evaluation. Distributed systems had been around for decades, and the Middleware conference itself resulted from the combination of three previous conferences. But the attempt to build common platforms for many di?erent applications requireda unique combinationofhi- level abstraction and low-level optimization, and presented challenges di?erent from building a monolithic distributed system. Since that ?rst conference, the notion of what constitutes "middleware" has changed somewhat, and the focus of research papers has changed with it. The ?rst edition focused heavily on distributed objects as a metaphor for building systems, including six papers with "CORBA" or "ORB" in the title. In f- lowing years, the conference broadened to cover publish/subscribe messaging, peer-to-peer systems, distributed databases, Web services, and automated m- agement, among other topics. Innovative techniques and architectures surfaced in workshops, and expanded to become themes of the main conference, while changes in the industry and advances in other research areas helped to shape research agendas. This tenth edition includes papers on next-generation pl- forms (such as stream systems, pervasive systems and cloud systems), managing enterprise data centers, and platforms for building other platforms, among o- ers.
Security and Dependability for Ambient Intelligence is the primary publication of the SERENITY approach, which provides security and dependability (S&D) solutions for dynamic, highly distributed, heterogeneous systems. The objective of SERENITY is to enhance the security and dependability of ambient intelligence systems by providing a framework supporting the automated integration, configuration, monitoring and adaptation of security and dependability mechanisms. An edited volume contributed by world leaders in the field, this book covers the problems that the highly dynamic and heterogeneous nature of ambient intelligence systems poses to security and dependability and presents solutions to these problems. Security and Dependability for Ambient Intelligence is designed for researchers and practitioners focusing on the dynamic integration, deployment and verification of security and dependability solutions in highly distributed systems incorporating ambient intelligence features. It is also suitable as a reference or secondary text for advanced-level students in computer science and computer or electrical engineering.
Real-Time Systems in Mechatronic Applications brings together in one place important contributions and up-to-date research results in this fast moving area. Real-Time Systems in Mechatronic Applications serves as an excellent reference, providing insight into some of the most challenging research issues in the field. |
You may like...
UC/OS-III - The Real-Time Kernel and the…
Jean J. Labrosse, Juan P. Benavides, …
Hardcover
R2,203
Discovery Miles 22 030
The Continuing Arms Race - Code-Reuse…
Per Larsen, Ahmad-Reza Sadeghi
Hardcover
R2,482
Discovery Miles 24 820
The Networknomicon, or SNMP Mastery
Michael W Lucas, Abdul Alhazred
Hardcover
BTEC Nationals Information Technology…
Jenny Phillips, Alan Jarvis, …
Paperback
R1,018
Discovery Miles 10 180
|