![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer hardware & operating systems > Operating systems & graphical user interfaces (GUIs) > General
Web caching and content delivery technologies provide the
infrastructure on which systems are built for the scalable
distribution of information. This proceedings of the eighth annual
workshop, captures a cross-section of the latest issues and
techniques of interest to network architects and researchers in
large-scale content delivery. Topics covered include the
distribution of streaming multimedia, edge caching and computation,
multicast, delivery of dynamic content, enterprise content
delivery, streaming proxies and servers, content transcoding,
replication and caching strategies, peer-to-peer content delivery,
and Web prefetching.
A detailed introduction to interdisciplinary application area of distributed systems, namely the computer support of individuals trying to solve a problem in cooperation with each other but not necessarily having identical work places or working times. The book is addressed to students of distributed systems, communications, information science and socio-organizational theory, as well as to users and developers of systems with group communication and cooperation as top priorities.
Based on research and industry experience, this book structures the issues pertaining to grid computing security into three main categories: architecture-related, infrastructure-related, and management-related issues. It discusses all three categories in detail, presents existing solutions, standards, and products, and pinpoints their shortcomings and open questions. Together with a brief introduction into grid computing in general and underlying security technologies, this book offers the first concise and detailed introduction to this important area, targeting professionals in the grid industry as well as students.
This year, the IFIP Working Conference on Distributed and Parallel Embedded Sys tems (DIPES 2008) is held as part of the IFIP World Computer Congress, held in Milan on September 7 10, 2008. The embedded systems world has a great deal of experience with parallel and distributed computing. Many embedded computing systems require the high performance that can be delivered by parallel computing. Parallel and distributed computing are often the only ways to deliver adequate real time performance at low power levels. This year's conference attracted 30 submissions, of which 21 were accepted. Prof. Jor ] g Henkel of the University of Karlsruhe graciously contributed a keynote address on embedded computing and reliability. We would like to thank all of the program committee members for their diligence. Wayne Wolf, Bernd Kleinjohann, and Lisa Kleinjohann Acknowledgements We would like to thank all people involved in the organization of the IFIP World Computer Congress 2008, especially the IPC Co Chairs Judith Bishop and Ivo De Lotto, the Organization Chair Giulio Occhini, as well as the Publications Chair John Impagliazzo. Further thanks go to the authors for their valuable contributions to DIPES 2008. Last but not least we would like to acknowledge the considerable amount of work and enthusiasm spent by our colleague Claudius Stern in preparing theproceedingsofDIPES2008. Hemadeitpossibletoproducethemintheircurrent professional and homogeneous style."
Operating systems kernels are central to the functioning of computers. Security of the overall system, as well as its reliability and responsiveness, depend upon the correct functioning of the kernel. This unique approach - presenting a formal specification of a kernel - starts with basic constructs and develops a set of kernels; proofs are included as part of the text.
Parallel and distributed computing is one of the foremost
technologies for shaping New Horizons of Parallel and Distributed Computing is a collection of self-contained chapters written by pioneering researchers to provide solutions for newly emerging problems in this field. This volume will not only provide novel ideas, work in progress and state-of-the-art techniques in the field, but will also stimulate future research activities in the area of parallel and distributed computing with applications. New Horizons of Parallel and Distributed Computing is intended for industry researchers and developers, as well as for academic researchers and advanced-level students in computer science and electrical engineering. A valuable reference work, it is also suitable as a textbook.
DAPSYS (International Conference on Distributed and Parallel Systems) is an international biannual conference series dedicated to all aspects of distributed and parallel computing. DAPSYS 2008, the 7th International Conference on Distributed and Parallel Systems was held in September 2008 in Hungary. Distributed and Parallel Systems: Desktop Grid Computing, based on DAPSYS 2008, presents original research, novel concepts and methods, and outstanding results. Contributors investigate parallel and distributed techniques, algorithms, models and applications; present innovative software tools, environments and middleware; focus on various aspects of grid computing; and introduce novel methods for development, deployment, testing and evaluation. This volume features a special focus on desktop grid computing as well. Designed for a professional audience composed of practitioners and researchers in industry, this book is also suitable for advanced-level students in computer science.
This book constitutes the refereed proceedings of the 15th International Conference on Principles of Distributed Systems, OPODIS 2011, held in Toulouse, France, in December 2011. The 26 revised papers presented in this volume were carefully reviewed and selected from 96 submissions. They represent the current state of the art of the research in the field of the design, analysis and development of distributed and real-time systems.
This essential resource for professionals and advanced students in security programming and system design introduces the foundations of programming systems security and the theory behind access control models, and addresses emerging access control mechanisms.
This book studies algorithmic issues associated with cooperative execution of multiple independent tasks by distributed computing agents including partitionable networks. It provides the most significant algorithmic solution developed and available today for do-all computing for distributed systems (including partitionable networks), and is the first monograph that deals with do-all computing for distributed systems. The book is structured to meet the needs of a professional audience composed of researchers and practitioners in industry. This volume is also suitable for graduate-level students in computer science.
"The Testing Network" presents an integrated approach to testing based on cutting-edge methodologies, processes and tools in today's IT context. It means complex network-centric applications to be tested in heterogeneous IT infrastructures and in multiple test environments (also geographically distributed). The added-value of this book is the in-depth explanation of all processes and relevant methodologies and tools to address this complexity. Main aspects of testing are explained using TD/QC - the world-leader test platform. This up-to-date know-how is based on real-life IT experiences gained in large-scale projects of companies operating worldwide. The book is abundantly illustrated to better show all technical aspects of modern testing in a national and international context. The author has a deep expertise by designing and giving testing training in large companies using the above-mentioned tools and processes. "The Testing Network" is a unique synthesis of core test topics applied in real-life.
This book constitutes the thoroughly refereed post-conference proceedings of the Second International ICST Conference on Sensor Systems and Software, S-Cube 2010, held in Miami, Florida, USA, in December 2010. The 17 revised full papers presented were carefully reviewed and selected and cover a wide range of topics including sensor application programming paradigms, novel sensor applications, sensor network middleware, trust security and privacy, wireless sensor network management and monitoring, and sensor application development support systems.
This book provides an in-depth description of event-based systems, covering topics ranging from local event matching and distributed event forwarding algorithms, through a practical discussion of software engineering issues raised by the event-based style, to state-of-the-art research in event-based systems like composite event detection and security. The authors offer a comprehensive overview, and show the power of event-based architectures in modern system design, encouraging professionals to exploit this technique in next generation large-scale distributed applications like information dissemination, network monitoring, enterprise application integration, or mobile systems.
This book constitutes the thoroughly refereed post-conference proceedings of the Third International ICST Conference on Forensic Applications and Techniques in Telecommunications, Information and Multimedia, E-Forensics 2010, held in Shanghai, China, in November 2010. The 32 revised full papers presented were carefully reviewed and selected from 42 submissions in total. These, along with 5 papers from a collocated workshop of E-Forensics Law, cover a wide range of topics including digital evidence handling, data carving, records tracing, device forensics, data tamper identification, and mobile device locating.
Intelligent Integration of Information presents a collection of chapters bringing the science of intelligent integration forward. The focus on integration defines tasks that increase the value of information when information from multiple sources is accessed, related, and combined. This contributed volume has also been published as a special double issue of the Journal of Intelligent Information Systems (JIIS), Volume 6:2/3.
Over the last decade, we have witnessed a growing dependency on information technologyresultingina wide rangeofnew opportunities. Clearly, ithas become almost impossible to imagine life without a personal computer or laptop, or without a cell phone. Social network sites (SNS) are competing with face-- face encounters and may even oust them. Most SNS-adepts have hundreds of "friends," happily sharing pictures and pro?les and endless chitchat. We are on the threshold of the Internet of Things, where every object will have its RFID-tag. This will not only e?ect companies, who will be able to optimize their production and delivery processes, but also end users, who will be able to enjoy many new applications, ranging from smart shopping, and smart fridges to geo-localized services. In the near future, elderly people will be able to stay longer at home due to clever health monitoring systems. The sky seems to be the limit However, we have also seen the other side of the coin: viruses, Trojan horses, breaches of privacy, identity theft, and other security threats. Our real and virtual worlds are becoming increasingly vulnerable to attack. In order to encouragesecurity researchby both academia and industry and to stimulate the dissemination of results, conferences need to be organized. With the 11th edition of the joint IFIP TC-6 TC-11 Conference on C- munications and Multimedia Security (CMS 2010), the organizers resumed the tradition of previous CMS conferences after a three-year recess.
As almost no other technology, embedded systems is an essential element of many innovations in automotive engineering. New functions and improvements of already existing functions, as well as the compliance with traffic regulations and customer requirements, have only become possible by the increasing use of electronic systems, especially in the fields of driving, safety, reliability, and functionality. Along with the functionalities that increase in number and have to cooperate, the complexity of the entire system will increase. Synergy effects resulting from distributed application functionalities via several electronic control devies, exchanging information through the network brings about more complex system architectures with many different sub-networks, operating with different velocities and different protocol implementations. To manage the increasing complexity of these systems, a deterministic behaviour of the control units and the communication network must be provided for, in particular when dealing with a distributed functionality. From Specification to Embedded Systems Application documents recent approaches and results presented at the International Embedded Systems Symposium (IESS 2005), which was held in August 2005 in Manaus (Brazil) and sponsored by the International Federation for Information Processing (IFIP). The topics which have been chosen for this working conference
are very timely: design methodology, modeling, specification,
software synthesis, power management, formal verification, testing,
network, communication systems, distributed control systems,
resource management and special aspects in system design.
This book intends to inculcate the innovative ideas for the scheduling aspect in distributed computing systems. Although the models in this book have been designed for distributed systems, the same information is applicable for any type of system. The book will dramatically improve the design and management of the processes for industry professionals. It deals exclusively with the scheduling aspect, which finds little space in other distributed operating system books. Structured for a professional audience composed of researchers and practitioners in industry, this book is also suitable as a reference for graduate-level students.
As the telecommunication industry introduces new sophisticated technologies, the nature of services and the volume of demands have changed. Indeed, a broad range of new services for users appear, combining voice, data, graphics, video, etc. This implies new planning issues. Fiber transmission systems that can carry large amounts of data on a few strands of wire were introduced. These systems have such a large bandwidth that the failure of even a single transmission link: in the network can create a severe service loss to customers. Therefore, a very high level of service reliability is becoming imperative for both system users and service providers. Since equipment failures and accidents cannot be avoided entirely, networks have to be designed so as to "survive" failures. This is done by judiciously installing spare capacity over the network so that all traffic interrupted by a failure may be diverted around that failure by way of this spare or reserve capacity. This of course translates into huge investments for network operators. Designing such survivable networks while minimizing spare capacity costs is, not surprisingly, a major concern of operating companies which gives rise to very difficult combinatorial problems. In order to make telecommunication networks survivable, one can essentially use two different strategies: protection or restoration. The protection approach preas signs spare capacity to protect each element of the network independently, while the restoration approach spreads the redundant capacity over the whole network and uses it as required in order to restore the disrupted traffic."
The Extreme Programming Pocket Guide covers XP assumptions, principles, events, artifacts, roles, and resources, and more. It concisely explains the relationships between the XP practices. If you want to adopt XP in stages, the Extreme Programming Pocket Guide will help you choose what to apply and when. Concise and easy to use, this handy pocket guide to XP is a must-have quick reference for anyone implementing a test-driven development environment.
A number of widely used contemporary processors have instruction-set extensions for improved performance in multi-media applications. The aim is to allow operations to proceed on multiple pixels each clock cycle. Such instruction-sets have been incorporated both in specialist DSPchips such as the Texas C62xx (Texas Instruments, 1998) and in general purpose CPU chips like the Intel IA32 (Intel, 2000) or the AMD K6 (Advanced Micro Devices, 1999). These instruction-set extensions are typically based on the Single Instruc tion-stream Multiple Data-stream (SIMD) model in which a single instruction causes the same mathematical operation to be carried out on several operands, or pairs of operands, at the same time. The level or parallelism supported ranges from two floating point operations, at a time on the AMD K6 architecture to 16 byte operations at a time on the Intel P4 architecture. Whereas processor architectures are moving towards greater levels of parallelism, the most widely used programming languages such as C, Java and Delphi are structured around a model of computation in which operations takeplace on a single value at a time. This was appropriate when processors worked this way, but has become an impediment to programmers seeking to make use of the performance offered by multi-media instruction -sets. The introduction of SIMD instruction sets (Peleg et al."
This book constitutes the thoroughly refereed joint post proceedings of two international workshops, the 5th International Workshop on Data Privacy Management, DPM 2010, and the 3rd International Workshop on Autonomous and Spontaneous Security, SETOP 2010, collocated with the ESORICS 2010 symposium in Athens, Greece, in September 2010. The 9 revised full papers for DPM 2010 presented together with two keynote talks are accompanied by 7 revised full papers of SETOP 2010; all papers were carefully reviewed and selected for inclusion in the book. The DPM 2010 papers cover topics such as how to translate the high-level business goals into system-level privacy policies, administration of privacy-sensitive data, privacy data integration and engineering, privacy access control mechanisms, information-oriented security, and query execution on privacy-sensitive data for partial answers. The SETOP 2010 papers address several specific aspects of the previously cited topics, as for instance the autonomic administration of security policies, secure P2P storage, RFID authentication, anonymity in reputation systems, etc.
This book constitutes the refereed proceedings of the Third
International Symposium on NASA Formal Methods, NFM 2011, held in
Pasadena, CA, USA, in April 2011.
As future generation information technology (FGIT) becomes specialized and fr- mented, it is easy to lose sight that many topics in FGIT have common threads and, because of this, advances in one discipline may be transmitted to others. Presentation of recent results obtained in different disciplines encourages this interchange for the advancement of FGIT as a whole. Of particular interest are hybrid solutions that c- bine ideas taken from multiple disciplines in order to achieve something more signi- cant than the sum of the individual parts. Through such hybrid philosophy, a new principle can be discovered, which has the propensity to propagate throughout mul- faceted disciplines. FGIT 2009 was the first mega-conference that attempted to follow the above idea of hybridization in FGIT in a form of multiple events related to particular disciplines of IT, conducted by separate scientific committees, but coordinated in order to expose the most important contributions. It included the following international conferences: Advanced Software Engineering and Its Applications (ASEA), Bio-Science and Bio-Technology (BSBT), Control and Automation (CA), Database Theory and Application (DTA), D- aster Recovery and Business Continuity (DRBC; published independently), Future G- eration Communication and Networking (FGCN) that was combined with Advanced Communication and Networking (ACN), Grid and Distributed Computing (GDC), M- timedia, Computer Graphics and Broadcasting (MulGraB), Security Technology (SecTech), Signal Processing, Image Processing and Pattern Recognition (SIP), and- and e-Service, Science and Technology (UNESST).
Explains fault tolerance in clear terms, with concrete examples
drawn from real-world settings |
You may like...
Nanotechnologies and Nanomaterials for…
Giuseppe Lazzara, Rawil F Fakhrullin
Paperback
Balancing of Linkages and Robot…
Vigen Arakelian, Sebastien Briot
Hardcover
Tibetan Buddhism and Mystical Experience
Yaroslav Komarovski
Hardcover
R3,567
Discovery Miles 35 670
Rail Vehicle Mechatronics
Maksym Spiryagin, Stefano Bruni, …
Hardcover
R4,508
Discovery Miles 45 080
The Legend Of Zola Mahobe - And The…
Don Lepati, Nikolaos Kirkinis
Paperback
(1)R480 Discovery Miles 4 800
|