Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > Computer hardware & operating systems > Computer architecture & logic design > Parallel processing
Edsger Wybe Dijkstra (1930-2002) was one of the most influential researchers in the history of computer science, making fundamental contributions to both the theory and practice of computing. Early in his career, he proposed the single-source shortest path algorithm, now commonly referred to as Dijkstra's algorithm. He wrote (with Jaap Zonneveld) the first ALGOL 60 compiler, and designed and implemented with his colleagues the influential THE operating system. Dijkstra invented the field of concurrent algorithms, with concepts such as mutual exclusion, deadlock detection, and synchronization. A prolific writer and forceful proponent of the concept of structured programming, he convincingly argued against the use of the Go To statement. In 1972 he was awarded the ACM Turing Award for "fundamental contributions to programming as a high, intellectual challenge; for eloquent insistence and practical demonstration that programs should be composed correctly, not just debugged into correctness; for illuminating perception of problems at the foundations of program design." Subsequently he invented the concept of self-stabilization relevant to fault-tolerant computing. He also devised an elegant language for nondeterministic programming and its weakest precondition semantics, featured in his influential 1976 book A Discipline of Programming in which he advocated the development of programs in concert with their correctness proofs. In the later stages of his life, he devoted much attention to the development and presentation of mathematical proofs, providing further support to his long-held view that the programming process should be viewed as a mathematical activity. In this unique new book, 31 computer scientists, including five recipients of the Turing Award, present and discuss Dijkstra's numerous contributions to computing science and assess their impact. Several authors knew Dijkstra as a friend, teacher, lecturer, or colleague. Their biographical essays and tributes provide a fascinating multi-author picture of Dijkstra, from the early days of his career up to the end of his life.
This book is a celebration of Leslie Lamport's work on concurrency, interwoven in four-and-a-half decades of an evolving industry: from the introduction of the first personal computer to an era when parallel and distributed multiprocessors are abundant. His works lay formal foundations for concurrent computations executed by interconnected computers. Some of the algorithms have become standard engineering practice for fault tolerant distributed computing - distributed systems that continue to function correctly despite failures of individual components. He also developed a substantial body of work on the formal specification and verification of concurrent systems, and has contributed to the development of automated tools applying these methods. Part I consists of technical chapters of the book and a biography. The technical chapters of this book present a retrospective on Lamport's original ideas from experts in the field. Through this lens, it portrays their long-lasting impact. The chapters cover timeless notions Lamport introduced: the Bakery algorithm, atomic shared registers and sequential consistency; causality and logical time; Byzantine Agreement; state machine replication and Paxos; temporal logic of actions (TLA). The professional biography tells of Lamport's career, providing the context in which his work arose and broke new grounds, and discusses LaTeX - perhaps Lamport's most influential contribution outside the field of concurrency. This chapter gives a voice to the people behind the achievements, notably Lamport himself, and additionally the colleagues around him, who inspired, collaborated, and helped him drive worldwide impact. Part II consists of a selection of Leslie Lamport's most influential papers. This book touches on a lifetime of contributions by Leslie Lamport to the field of concurrency and on the extensive influence he had on people working in the field. It will be of value to historians of science, and to researchers and students who work in the area of concurrency and who are interested to read about the work of one of the most influential researchers in this field.
In recent years, most applications deal with constraint decision-making systems as problems are based on imprecise information and parameters. It is difficult to understand the nature of data based on applications and it requires a specific model for understanding the nature of the system. Further research on constraint decision-making systems in engineering is required. Constraint Decision-Making Systems in Engineering derives and explores several types of constraint decisions in engineering and focuses on new and innovative conclusions based on problems, robust and efficient systems, and linear and non-linear applications. Covering topics such as fault detection, data mining techniques, and knowledge-based management, this premier reference source is an essential resource for engineers, managers, computer scientists, students and educators of higher education, librarians, researchers, and academicians.
Distributed systems intertwine with our everyday lives. The benefits and current shortcomings of the underpinning technologies are experienced by a wide range of people and their smart devices. With the rise of large-scale IoT and similar distributed systems, cloud bursting technologies, and partial outsourcing solutions, private entities are encouraged to increase their efficiency and offer unparalleled availability and reliability to their users. Applying Integration Techniques and Methods in Distributed Systems is a critical scholarly publication that defines the current state of distributed systems, determines further goals, and presents architectures and service frameworks to achieve highly integrated distributed systems and presents solutions to integration and efficient management challenges faced by current and future distributed systems. Highlighting topics such as multimedia, programming languages, and smart environments, this book is ideal for system administrators, integrators, designers, developers, researchers, and academicians.
Recent years have witnessed the rise of analysis of real-world massive and complex phenomena in graphs; to efficiently solve these large-scale graph problems, it is necessary to exploit high performance computing (HPC), which accelerates the innovation process for discovery and invention of new products and procedures in network science. Creativity in Load-Balance Schemes for Multi/Many-Core Heterogeneous Graph Computing: Emerging Research and Opportunities is a critical scholarly resource that examines trends, challenges, and collaborative processes in emerging fields within complex network analysis. Featuring coverage on a broad range of topics such as high-performance computing, big data, network science, and accelerated network traversal, this book is geared towards data analysts, researchers, students in information communication technology (ICT), program developers, and academics.
Present day sophisticated, adaptive, and autonomous (to a certain degree) robotic technology is a radically new stimulus for the cognitive system of the human learner from the earliest to the oldest age. It deserves extensive, thorough, and systematic research based on novel frameworks for analysis, modelling, synthesis, and implementation of CPSs for social applications. Cyber-Physical Systems for Social Applications is a critical scholarly book that examines the latest empirical findings for designing cyber-physical systems for social applications and aims at forwarding the symbolic human-robot perspective in areas that include education, social communication, entertainment, and artistic performance. Highlighting topics such as evolinguistics, human-robot interaction, and neuroinformatics, this book is ideally designed for social network developers, cognitive scientists, education science experts, evolutionary linguists, researchers, and academicians.
As the future of software development in a global environment continues to be influenced by the areas of service oriented architecture (SOA) and cloud computing, many legacy applications will need to migrate these environments to take advantage of the benefits offered by the service environment. Migrating Legacy Applications: Challenges in Service Oriented Architecture and Cloud Computing Environments presents a closer look at the partnership between service oriented architecture and cloud computing environments while analyzing potential solutions to challenges related to the migration of legacy applications. This reference is essential for students and university scholars alike.
As software and computer hardware grows in complexity, networks have grown to match. The increasing scale, complexity, heterogeneity, and dynamism of communication networks, resources, and applications has made distributed computing systems brittle, unmanageable, and insecure. Internet and Distributed Computing Advancements: Theoretical Frameworks and Practical Applications is a vital compendium of chapters on the latest research within the field of distributed computing, capturing trends in the design and development of Internet and distributed computing systems that leverage autonomic principles and techniques. The chapters provided within this collection offer a holistic approach for the development of systems that can adapt themselves to meet requirements of performance, fault tolerance, reliability, security, and Quality of Service (QoS) without manual intervention.
This book precisely formulates and simplifies the presentation of Instruction Level Parallelism (ILP) compilation techniques. It uniquely offers consistent and uniform descriptions of the code transformations involved. Due to the ubiquitous nature of ILP in virtually every processor built today, from general purpose CPUs to application-specific and embedded processors, this book is useful to the student, the practitioner and also the researcher of advanced compilation techniques. With an emphasis on fine-grain instruction level parallelism, this book will also prove interesting to researchers and students of parallelism at large, in as much as the techniques described yield insights that go beyond superscalar and VLIW (Very Long Instruction Word) machines compilation and are more widely applicable to optimizing compilers in general. ILP techniques have found wide and crucial application in Design Automation, where they have been used extensively in the optimization of performance as well as area and power minimization of computer designs.
Massively Parallel Systems (MPSs) with their scalable computation and storage space promises are becoming increasingly important for high-performance computing. The growing acceptance of MPSs in academia is clearly apparent. However, in industrial companies, their usage remains low. The programming of MPSs is still the big obstacle, and solving this software problem is sometimes referred to as one of the most challenging tasks of the 1990's. The 1994 working conference on "Programming Environments for Massively Parallel Systems" was the latest event of the working group WG 10.3 of the International Federation for Information Processing (IFIP) in this field. It succeeded the 1992 conference in Edinburgh on "Programming Environments for Parallel Computing." The research and development work discussed at the conference addresses the entire spectrum of software problems including virtual machines which are less cumbersome to program; more convenient programming models; advanced programming languages, and especially more sophisticated programming tools; but also algorithms and applications.
This volume gives an overview of the state-of-the-art with respect to the development of all types of parallel computers and their application to a wide range of problem areas.
In the past two decades, breakthroughs in computer technology have made a tremendous impact on optimization. In particular, availability of parallel computers has created substantial interest in exploring the use of parallel processing for solving discrete and global optimization problems. The chapters in this volume cover a broad spectrum of recent research in parallel processing of discrete and related problems. The topics discussed include distributed branch-and-bound algorithms, parallel genetic algorithms for large scale discrete problems, simulated annealing, parallel branch-and-bound search under limited-memory constraints, parallelization of greedy randomized adaptive search procedures, parallel optical models of computing, randomized parallel algorithms, general techniques for the design of parallel discrete algorithms, parallel algorithms for the solution of quadratic assignment and satisfiability problems. The book will be a valuable source of information to faculty, students and researchers in combinatorial optimization and related areas.
The book "Parallel Computing" deals with the topics of current interest in high performance computing, viz. pipeline and parallel processing architectures, and the whole book is based on treatment of these ideas. The present revised edition is updated with the addition of topics like processor performance and technology developments in chapter 1 and advanced pipeline processing on today's high performance processors in chapter 2. A new chapter on neurocomputing and two new sections on Branch prediction and scoreboard are the other major changes done to make the book more viable.
A number of widely used contemporary processors have instruction-set extensions for improved performance in multi-media applications. The aim is to allow operations to proceed on multiple pixels each clock cycle. Such instruction-sets have been incorporated both in specialist DSPchips such as the Texas C62xx (Texas Instruments, 1998) and in general purpose CPU chips like the Intel IA32 (Intel, 2000) or the AMD K6 (Advanced Micro Devices, 1999). These instruction-set extensions are typically based on the Single Instruc tion-stream Multiple Data-stream (SIMD) model in which a single instruction causes the same mathematical operation to be carried out on several operands, or pairs of operands, at the same time. The level or parallelism supported ranges from two floating point operations, at a time on the AMD K6 architecture to 16 byte operations at a time on the Intel P4 architecture. Whereas processor architectures are moving towards greater levels of parallelism, the most widely used programming languages such as C, Java and Delphi are structured around a model of computation in which operations takeplace on a single value at a time. This was appropriate when processors worked this way, but has become an impediment to programmers seeking to make use of the performance offered by multi-media instruction -sets. The introduction of SIMD instruction sets (Peleg et al."
This book offers readers a set of new approaches and tools a set of tools and techniques for facing challenges in parallelization with design of embedded systems. It provides an advanced parallel simulation infrastructure for efficient and effective system-level model validation and development so as to build better products in less time. Since parallel discrete event simulation (PDES) has the potential to exploit the underlying parallel computational capability in today's multi-core simulation hosts, the author begins by reviewing the parallelization of discrete event simulation, identifying problems and solutions. She then describes out-of-order parallel discrete event simulation (OoO PDES), a novel approach for efficient validation of system-level designs by aggressively exploiting the parallel capabilities of todays' multi-core PCs. This approach enables readers to design simulators that can fully exploit the parallel processing capability of the multi-core system to achieve fast speed simulation, without loss of simulation and timing accuracy. Based on this parallel simulation infrastructure, the author further describes automatic approaches that help the designer quickly to narrow down the debugging targets in faulty ESL models with parallelism. |
You may like...
Advances in High Performance Computing…
Lucio Grandinetti, Etc
Hardcover
R2,520
Discovery Miles 25 200
Distributed and Parallel Systems - From…
Peter Kacsuk, Gabriele Kotsis
Hardcover
R5,568
Discovery Miles 55 680
Parallel and Distributed Information…
Jeffrey F. Naughton, Gerhard Weikum
Hardcover
R2,849
Discovery Miles 28 490
Parallel Computing on Distributed Memory…
Fusun Oezguner, Fikret Ercal
Hardcover
R2,988
Discovery Miles 29 880
Algorithms for Parallel Processing
Michael T. Heath, Abhiram Ranade, …
Hardcover
R3,012
Discovery Miles 30 120
|