0
Your cart

Your cart is empty

Browse All Departments
Price
  • R100 - R250 (6)
  • R250 - R500 (17)
  • R500+ (475)
  • -
Status
Format
Author / Contributor
Publisher

Books > Computing & IT > Computer hardware & operating systems > Computer architecture & logic design > Parallel processing

Parallel Computing (Hardcover): Moreshwar R. Bhujade Parallel Computing (Hardcover)
Moreshwar R. Bhujade
R1,040 Discovery Miles 10 400 Ships in 12 - 17 working days

The book "Parallel Computing" deals with the topics of current interest in high performance computing, viz. pipeline and parallel processing architectures, and the whole book is based on treatment of these ideas. The present revised edition is updated with the addition of topics like processor performance and technology developments in chapter 1 and advanced pipeline processing on today's high performance processors in chapter 2. A new chapter on neurocomputing and two new sections on Branch prediction and scoreboard are the other major changes done to make the book more viable.

SIMD Programming Manual for Linux and Windows (Hardcover, 2004 ed.): Paul Cockshott, Kenneth Renfrew SIMD Programming Manual for Linux and Windows (Hardcover, 2004 ed.)
Paul Cockshott, Kenneth Renfrew
R3,244 Discovery Miles 32 440 Ships in 12 - 17 working days

A number of widely used contemporary processors have instruction-set extensions for improved performance in multi-media applications. The aim is to allow operations to proceed on multiple pixels each clock cycle. Such instruction-sets have been incorporated both in specialist DSPchips such as the Texas C62xx (Texas Instruments, 1998) and in general purpose CPU chips like the Intel IA32 (Intel, 2000) or the AMD K6 (Advanced Micro Devices, 1999). These instruction-set extensions are typically based on the Single Instruc tion-stream Multiple Data-stream (SIMD) model in which a single instruction causes the same mathematical operation to be carried out on several operands, or pairs of operands, at the same time. The level or parallelism supported ranges from two floating point operations, at a time on the AMD K6 architecture to 16 byte operations at a time on the Intel P4 architecture. Whereas processor architectures are moving towards greater levels of parallelism, the most widely used programming languages such as C, Java and Delphi are structured around a model of computation in which operations takeplace on a single value at a time. This was appropriate when processors worked this way, but has become an impediment to programmers seeking to make use of the performance offered by multi-media instruction -sets. The introduction of SIMD instruction sets (Peleg et al."

Out-of-order Parallel Discrete Event Simulation for Electronic System-level Design (Hardcover, 2015 ed.): Weiwei Chen Out-of-order Parallel Discrete Event Simulation for Electronic System-level Design (Hardcover, 2015 ed.)
Weiwei Chen
R2,921 Discovery Miles 29 210 Ships in 10 - 15 working days

This book offers readers a set of new approaches and tools a set of tools and techniques for facing challenges in parallelization with design of embedded systems. It provides an advanced parallel simulation infrastructure for efficient and effective system-level model validation and development so as to build better products in less time. Since parallel discrete event simulation (PDES) has the potential to exploit the underlying parallel computational capability in today's multi-core simulation hosts, the author begins by reviewing the parallelization of discrete event simulation, identifying problems and solutions. She then describes out-of-order parallel discrete event simulation (OoO PDES), a novel approach for efficient validation of system-level designs by aggressively exploiting the parallel capabilities of todays' multi-core PCs. This approach enables readers to design simulators that can fully exploit the parallel processing capability of the multi-core system to achieve fast speed simulation, without loss of simulation and timing accuracy. Based on this parallel simulation infrastructure, the author further describes automatic approaches that help the designer quickly to narrow down the debugging targets in faulty ESL models with parallelism.

Specification and Proof in Real Time CSP (Hardcover): Jim Davies Specification and Proof in Real Time CSP (Hardcover)
Jim Davies
R3,063 Discovery Miles 30 630 Ships in 12 - 17 working days

Computing systems are becoming highly complex, harder to understand, and therefore more prone to failure. Where such systems control aircraft for example, system failure could have disastrous consequences. It is important therefore that we are able to employ mathematical techniques to specify the behavior of critical systems. This thesis uses the theory of Communicating Sequential Processes to show how a real-time system (a system that maintains a continuous interaction with its environment) may be specified. Included is a case study in which a local area network protocol is described at two levels of abstraction, and a general method for structuring CSP descriptions of layered protocols is given. The research contained here represents the very latest work on the specification and verification of real-time systems.

Advances in High Performance Computing - Proceedings of the NATO Advanced Research Workshop on High Performance Computing -... Advances in High Performance Computing - Proceedings of the NATO Advanced Research Workshop on High Performance Computing - Technology and Applications, Cetraro, Italy, 24-26 June 1996 (Hardcover)
Lucio Grandinetti, Etc
R2,641 Discovery Miles 26 410 Ships in 12 - 17 working days

The book discusses some key scientific and technological developments in high performance computing, identifies significant trends, and defines desirable research objectives. It covers general concepts and emerging systems, software technology, algorithms and applications. Coverage includes hardware, software tools, networks and numerical methods, new computer architectures, and a discussion of future trends. Beyond purely scientific/engineering computing, the book extends to coverage of enterprise-wide, commercial applications, including papers on performance and scalability of database servers and Oracle DBM systems. Audience: Most papers are research level, but some are suitable for computer literate managers and technicians, making the book useful to users of commercial parallel computers.

The Sourcebook of Parallel Computing (Hardcover, New): Jack Dongarra, Ian Foster, Geoffrey C. Fox, William Gropp, Ken Kennedy,... The Sourcebook of Parallel Computing (Hardcover, New)
Jack Dongarra, Ian Foster, Geoffrey C. Fox, William Gropp, Ken Kennedy, …
R2,049 Discovery Miles 20 490 Ships in 12 - 17 working days


Parallel Computing is a compelling vision of how computation can seamlessly scale from a single processor to virtually limitless computing power. Unfortunately, the scaling of application performance has not matched peak speed, and the programming burden for these machines remains heavy. The applications must be programmed to exploit parallelism in the most efficient way possible. Today, the responsibility for achieving the vision of scalable parallelism remains in the hands of the application developer.


This book represents the collected knowledge and experience of over 60 leading parallel computing researchers. They offer students, scientists and engineers a complete sourcebook with solid coverage of parallel computing hardware, programming considerations, algorithms, software and enabling technologies, as well as several parallel application case studies. The Sourcebook of Parallel Computing offers extensive tutorials and detailed documentation of the advanced strategies produced by research over the last two decades
application case studies. The Sourcebook of Parallel Computing offers extensive tutorials and detailed documentation of the advanced strategies produced by research over the last two decades
* Provides a solid background in parallel computing technologies
* Examines the technologies available and teaches students and practitioners how to select and apply them
* Presents case studies in a range of application areas including Chemistry, Image Processing, Data Mining, Ocean Modeling and Earthquake Simulation
* Considers the future development of parallel computing technologies and the kinds of applications they will support

Distributed and Parallel Systems - From Instruction Parallelism to Cluster Computing (Hardcover, 2000 ed.): Peter Kacsuk,... Distributed and Parallel Systems - From Instruction Parallelism to Cluster Computing (Hardcover, 2000 ed.)
Peter Kacsuk, Gabriele Kotsis
R5,847 Discovery Miles 58 470 Ships in 10 - 15 working days

Distributed and Parallel Systems: From Instruction Parallelism to Cluster Computing is the proceedings of the third Austrian-Hungarian Workshop on Distributed and Parallel Systems organized jointly by the Austrian Computer Society and the MTA SZTAKI Computer and Automation Research Institute. This book contains 18 full papers and 12 short papers from 14 countries around the world, including Japan, Korea and Brazil. The paper sessions cover a broad range of research topics in the area of parallel and distributed systems, including software development environments, performance evaluation, architectures, languages, algorithms, web and cluster computing. This volume will be useful to researchers and scholars interested in all areas related to parallel and distributed computing systems.

Parallel Computing Using Optical Interconnections (Hardcover, 1998 ed.): Keqin Li, Yi Pan, Si Qing Zheng Parallel Computing Using Optical Interconnections (Hardcover, 1998 ed.)
Keqin Li, Yi Pan, Si Qing Zheng
R4,535 Discovery Miles 45 350 Ships in 12 - 17 working days

Advances in optical technologies have made it possible to implement optical interconnections in future massively parallel processing systems. Photons are non-charged particles, and do not naturally interact. Consequently, there are many desirable characteristics of optical interconnects, e.g. high speed (speed of light), increased fanout, high bandwidth, high reliability, longer interconnection lengths, low power requirements, and immunity to EMI with reduced crosstalk. Optics can utilize free-space interconnects as well as guided wave technology, neither of which has the problems of VLSI technology mentioned above. Optical interconnections can be built at various levels, providing chip-to-chip, module-to-module, board-to-board, and node-to-node communications. Massively parallel processing using optical interconnections poses new challenges; new system configurations need to be designed, scheduling and data communication schemes based on new resource metrics need to be investigated, algorithms for a wide variety of applications need to be developed under the novel computation models that optical interconnections permit, and so on. Parallel Computing Using Optical Interconnections is a collection of survey articles written by leading and active scientists in the area of parallel computing using optical interconnections. This is the first book which provides current and comprehensive coverage of the field, reflects the state of the art from high-level architecture design and algorithmic points of view, and points out directions for further research and development.

Cellular Automata - A Parallel Model (Hardcover, 1999 ed.): M. Delorme, J. Mazoyer Cellular Automata - A Parallel Model (Hardcover, 1999 ed.)
M. Delorme, J. Mazoyer
R4,553 Discovery Miles 45 530 Ships in 12 - 17 working days

Cellular automata can be viewed both as computational models and modelling systems of real processes. This volume emphasises the first aspect. In articles written by leading researchers, sophisticated massive parallel algorithms (firing squad, life, Fischer's primes recognition) are treated. Their computational power and the specific complexity classes they determine are surveyed, while some recent results in relation to chaos from a new dynamic systems point of view are also presented. Audience: This book will be of interest to specialists of theoretical computer science and the parallelism challenge.

Parallel Computer Architecture - A Hardware/Software Approach (Hardcover): David Culler, J.P. Singh, Anoop Gupta Parallel Computer Architecture - A Hardware/Software Approach (Hardcover)
David Culler, J.P. Singh, Anoop Gupta
R2,905 Discovery Miles 29 050 Ships in 12 - 17 working days

The most exciting development in parallel computer architecture is the convergence of traditionally disparate approaches on a common machine structure. This book explains the forces behind this convergence of shared-memory, message-passing, data parallel, and data-driven computing architectures. It then examines the design issues that are critical to all parallel architecture across the full range of modern design, covering data access, communication performance, coordination of cooperative work, and correct implementation of useful semantics. It not only describes the hardware and software techniques for addressing each of these issues but also explores how these techniques interact in the same system. Examining architecture from an application-driven perspective, it provides comprehensive discussions of parallel programming for high performance and of workload-driven evaluation, based on understanding hardware-software interactions.
* synthesizes a decade of research and development for practicing engineers, graduate students, and researchers in parallel computer architecture, system software, and applications development
* presents in-depth application case studies from computer graphics, computational science and engineering, and data mining to demonstrate sound quantitative evaluation of design trade-offs
* describes the process of programming for performance, including both the architecture-independent and architecture-dependent aspects, with examples and case-studies
* illustrates bus-based and network-based parallel systems with case studies of more than a dozen important commercial designs

Development and Deployment of Multiplayer Online Games, Vol. II - DIY, (Re)Actors, Client Arch., Unity/UE4/ Lumberyard/Urho3D... Development and Deployment of Multiplayer Online Games, Vol. II - DIY, (Re)Actors, Client Arch., Unity/UE4/ Lumberyard/Urho3D (Hardcover)
'No Bugs' Hare; Contributions by Sergey Ignatchenko
R922 Discovery Miles 9 220 Ships in 12 - 17 working days
Parallel Numerical Computation with Applications (Hardcover, 1999 ed.): Laurence Tianruo Yang Parallel Numerical Computation with Applications (Hardcover, 1999 ed.)
Laurence Tianruo Yang
R4,521 Discovery Miles 45 210 Ships in 12 - 17 working days

Parallel Numerical Computations with Applications contains selected edited papers presented at the 1998 Frontiers of Parallel Numerical Computations and Applications Workshop, along with invited papers from leading researchers around the world. These papers cover a broad spectrum of topics on parallel numerical computation with applications; such as advanced parallel numerical and computational optimization methods, novel parallel computing techniques, numerical fluid mechanics, and other applications related to material sciences, signal and image processing, semiconductor technology, and electronic circuits and systems design. This state-of-the-art volume will be an up-to-date resource for researchers in the areas of parallel and distributed computing.

Parallel and Distributed Information Systems (Hardcover, Reprinted From Parallel And Distributed Databases,): Jeffrey F.... Parallel and Distributed Information Systems (Hardcover, Reprinted From Parallel And Distributed Databases,)
Jeffrey F. Naughton, Gerhard Weikum
R2,989 Discovery Miles 29 890 Ships in 10 - 15 working days

Parallel and Distributed Information Systems brings together in one place important contributions and up-to-date research results in this fast moving area. Parallel and Distributed Information Systems serves as an excellent reference, providing insight into some of the most challenging research issues in the field.

Concurrent Scientific Computing (Hardcover, 1994 ed.): Eric F. Van De Velde Concurrent Scientific Computing (Hardcover, 1994 ed.)
Eric F. Van De Velde
R2,057 Discovery Miles 20 570 Ships in 12 - 17 working days

Mathematics is playing an ever more important role in the physical and biological sciences, provoking a blurring of boundaries between scientific dis ciplines and a resurgence of interest in the modern as well as the classical techniques of applied mathematics. This renewal of interest, both in research and teaching, has led to the establishment of the series: Texts in Applied Mathe matics (TAM). The development of new courses is a natural consequence of a high level of excitement on the research frontier as newer techniques, such as numerical and symbolic computer systems, dynamical systems, and chaos, mix with and reinforce the traditional methods of applied mathematics. Thus, the purpose of this textbook series is to meet the current and future needs of these advances and encourage the teaching of new courses. TAM will publish textbooks suitable for use in advanced undergraduate and beginning graduate courses, and will complement the Applied Mathematical Sciences (AMS) series, which will focus on advanced textbooks and research level monographs. Preface A successful concurrent numerical simulation requires physics and math ematics to develop and analyze the model, numerical analysis to develop solution methods, and computer science to develop a concurrent implemen tation. No single course can or should cover all these disciplines. Instead, this course on concurrent scientific computing focuses on a topic that is not covered or is insufficiently covered by other disciplines: the algorith mic structure of numerical methods."

Solving Partial Differential Equations On Parallel Computers (Hardcover): Jianping Zhu Solving Partial Differential Equations On Parallel Computers (Hardcover)
Jianping Zhu
R2,601 Discovery Miles 26 010 Ships in 12 - 17 working days

This is an introductory book on supercomputer applications written by a researcher who is working on solving scientific and engineering application problems on parallel computers. The book is intended to quickly bring researchers and graduate students working on numerical solutions of partial differential equations with various applications into the area of parallel processing.The book starts from the basic concepts of parallel processing, like speedup, efficiency and different parallel architectures, then introduces the most frequently used algorithms for solving PDEs on parallel computers, with practical examples. Finally, it discusses more advanced topics, including different scalability metrics, parallel time stepping algorithms and new architectures and heterogeneous computing networks which have emerged in the last few years of high performance computing. Hundreds of references are also included in the book to direct interested readers to more detailed and in-depth discussions of specific topics.

Parallel Computing on Distributed Memory Multiprocessors (Hardcover, 1993 ed.): Fusun Oezguner, Fikret Ercal Parallel Computing on Distributed Memory Multiprocessors (Hardcover, 1993 ed.)
Fusun Oezguner, Fikret Ercal
R3,132 Discovery Miles 31 320 Ships in 10 - 15 working days

Advances in microelectronic technology have made massively parallel computing a reality and triggered an outburst of research activity in parallel processing architectures and algorithms. Distributed memory multiprocessors - parallel computers that consist of microprocessors connected in a regular topology - are increasingly being used to solve large problems in many application areas. In order to use these computers for a specific application, existing algorithms need to be restructured for the architecture and new algorithms developed. The performance of a computation on a distributed memory multiprocessor is affected by the node and communication architecture, the interconnection network topology, the I/O subsystem, and the parallel algorithm and communication protocols. Each of these parametersis a complex problem, and solutions require an understanding of the interactions among them. This book is based on the papers presented at the NATO Advanced Study Institute held at Bilkent University, Turkey, in July 1991. The book is organized in five parts: Parallel computing structures and communication, Parallel numerical algorithms, Parallel programming, Fault tolerance, and Applications and algorithms.

Algorithms for Parallel Processing (Hardcover, 1999 ed.): Michael T. Heath, Abhiram Ranade, Robert S. Schreiber Algorithms for Parallel Processing (Hardcover, 1999 ed.)
Michael T. Heath, Abhiram Ranade, Robert S. Schreiber
R3,156 Discovery Miles 31 560 Ships in 10 - 15 working days

This IMA Volume in Mathematics and its Applications ALGORITHMS FOR PARALLEL PROCESSING is based on the proceedings of a workshop that was an integral part of the 1996-97 IMA program on "MATHEMATICS IN HIGH-PERFORMANCE COMPUTING. " The workshop brought together algorithm developers from theory, combinatorics, and scientific computing. The topics ranged over models, linear algebra, sorting, randomization, and graph algorithms and their analysis. We thank Michael T. Heath of University of lllinois at Urbana (Com puter Science), Abhiram Ranade of the Indian Institute of Technology (Computer Science and Engineering), and Robert S. Schreiber of Hewlett Packard Laboratories for their excellent work in organizing the workshop and editing the proceedings. We also take this opportunity to thank the National Science Founda tion (NSF) and the Army Research Office (ARO), whose financial support made the workshop possible. A vner Friedman Robert Gulliver v PREFACE The Workshop on Algorithms for Parallel Processing was held at the IMA September 16 - 20, 1996; it was the first workshop of the IMA year dedicated to the mathematics of high performance computing. The work shop organizers were Abhiram Ranade of The Indian Institute of Tech nology, Bombay, Michael Heath of the University of Illinois, and Robert Schreiber of Hewlett Packard Laboratories. Our idea was to bring together researchers who do innovative, exciting, parallel algorithms research on a wide range of topics, and by sharing insights, problems, tools, and methods to learn something of value from one another."

Parallel Computing in Optimization (Hardcover, 1997 ed.): A. Migdalas, Panos M. Pardalos, Sverre Storoy Parallel Computing in Optimization (Hardcover, 1997 ed.)
A. Migdalas, Panos M. Pardalos, Sverre Storoy
R8,639 Discovery Miles 86 390 Ships in 12 - 17 working days

During the last three decades, breakthroughs in computer technology have made a tremendous impact on optimization. In particular, parallel computing has made it possible to solve larger and computationally more difficult problems. The book covers recent developments in novel programming and algorithmic aspects of parallel computing as well as technical advances in parallel optimization. Each contribution is essentially expository in nature, but of scholarly treatment. In addition, each chapter includes a collection of carefully selected problems. The first two chapters discuss theoretical models for parallel algorithm design and their complexity. The next chapter gives the perspective of the programmer practicing parallel algorithm development on real world platforms. Solving systems of linear equations efficiently is of great importance not only because they arise in many scientific and engineering applications but also because algorithms for solving many optimization problems need to call system solvers and subroutines (chapters four and five). Chapters six through thirteen are dedicated to optimization problems and methods. They include parallel algorithms for network problems, parallel branch and bound techniques, parallel heuristics for discrete and continuous problems, decomposition methods, parallel algorithms for variational inequality problems, parallel algorithms for stochastic programming, and neural networks. Audience: Parallel Computing in Optimization is addressed not only to researchers of mathematical programming, but to all scientists in various disciplines who use optimization methods in parallel and multiprocessing environments to model and solve problems.

Software-Implemented Hardware Fault Tolerance (Hardcover): Olga Goloubeva, Maurizio Rebaudengo, Matteo Sonza Reorda, Massimo... Software-Implemented Hardware Fault Tolerance (Hardcover)
Olga Goloubeva, Maurizio Rebaudengo, Matteo Sonza Reorda, Massimo Violante
R4,519 Discovery Miles 45 190 Ships in 12 - 17 working days

This book presents the theory behind software-implemented hardware fault tolerance, as well as the practical aspects needed to put it to work on real examples. By evaluating accurately the advantages and disadvantages of the already available approaches, the book provides a guide to developers willing to adopt software-implemented hardware fault tolerance in their applications. Moreover, the book identifies open issues for researchers willing to improve the already available techniques.

Hierarchical Scheduling in Parallel and Cluster Systems (Hardcover, 2003 ed.): Sivarama Dandamudi Hierarchical Scheduling in Parallel and Cluster Systems (Hardcover, 2003 ed.)
Sivarama Dandamudi
R4,528 Discovery Miles 45 280 Ships in 12 - 17 working days

Multiple processor systems are an important class of parallel systems. Over the years, several architectures have been proposed to build such systems to satisfy the requirements of high performance computing. These architectures span a wide variety of system types. At the low end of the spectrum, we can build a small, shared-memory parallel system with tens of processors. These systems typically use a bus to interconnect the processors and memory. Such systems, for example, are becoming commonplace in high-performance graph ics workstations. These systems are called uniform memory access (UMA) multiprocessors because they provide uniform access of memory to all pro cessors. These systems provide a single address space, which is preferred by programmers. This architecture, however, cannot be extended even to medium systems with hundreds of processors due to bus bandwidth limitations. To scale systems to medium range i. e., to hundreds of processors, non-bus interconnection networks have been proposed. These systems, for example, use a multistage dynamic interconnection network. Such systems also provide global, shared memory like the UMA systems. However, they introduce local and remote memories, which lead to non-uniform memory access (NUMA) architecture. Distributed-memory architecture is used for systems with thousands of pro cessors. These systems differ from the shared-memory architectures in that there is no globally accessible shared memory. Instead, they use message pass ing to facilitate communication among the processors. As a result, they do not provide single address space."

Introduction to Parallel Processing - Algorithms and Architectures (Hardcover, 1999 ed.): Behrooz Parhami Introduction to Parallel Processing - Algorithms and Architectures (Hardcover, 1999 ed.)
Behrooz Parhami
R6,177 Discovery Miles 61 770 Ships in 10 - 15 working days

THE CONTEXT OF PARALLEL PROCESSING The field of digital computer architecture has grown explosively in the past two decades. Through a steady stream of experimental research, tool-building efforts, and theoretical studies, the design of an instruction-set architecture, once considered an art, has been transformed into one of the most quantitative branches of computer technology. At the same time, better understanding of various forms of concurrency, from standard pipelining to massive parallelism, and invention of architectural structures to support a reasonably efficient and user-friendly programming model for such systems, has allowed hardware performance to continue its exponential growth. This trend is expected to continue in the near future. This explosive growth, linked with the expectation that performance will continue its exponential rise with each new generation of hardware and that (in stark contrast to software) computer hardware will function correctly as soon as it comes off the assembly line, has its down side. It has led to unprecedented hardware complexity and almost intolerable dev- opment costs. The challenge facing current and future computer designers is to institute simplicity where we now have complexity; to use fundamental theories being developed in this area to gain performance and ease-of-use benefits from simpler circuits; to understand the interplay between technological capabilities and limitations, on the one hand, and design decisions based on user and application requirements on the other.

Tools and Environments for Parallel and Distributed Systems (Hardcover, 1996 ed.): Amr Zaky, Ted Lewis Tools and Environments for Parallel and Distributed Systems (Hardcover, 1996 ed.)
Amr Zaky, Ted Lewis
R4,537 Discovery Miles 45 370 Ships in 12 - 17 working days

Developing correct and efficient software is far more complex for parallel and distributed systems than it is for sequential processors. Some of the reasons for this added complexity are: the lack of a universally acceptable parallel and distributed programming paradigm, the criticality of achieving high performance, and the difficulty of writing correct parallel and distributed programs. These factors collectively influence the current status of parallel and distributed software development tools efforts. Tools and Environments for Parallel and Distributed Systems addresses the above issues by describing working tools and environments, and gives a solid overview of some of the fundamental research being done worldwide. Topics covered in this collection are: mainstream program development tools, performance prediction tools and studies; debugging tools and research; and nontraditional tools. Audience: Suitable as a secondary text for graduate level courses in software engineering and parallel and distributed systems, and as a reference for researchers and practitioners in industry.

Multithreaded Computer Architecture: A Summary of the State of the ART (Hardcover, 1994 ed.): Robert A. Iannucci, Guang R. Gao,... Multithreaded Computer Architecture: A Summary of the State of the ART (Hardcover, 1994 ed.)
Robert A. Iannucci, Guang R. Gao, Robert H. Halstead Jr, Burton Smith
R5,951 Discovery Miles 59 510 Ships in 10 - 15 working days

Multithreaded computer architecture has emerged as one of the most promising and exciting avenues for the exploitation of parallelism. This new field represents the confluence of several independent research directions which have united over a common set of issues and techniques. Multithreading draws on recent advances in dataflow, RISC, compiling for fine-grained parallel execution, and dynamic resource management. It offers the hope of dramatic performance increases through parallel execution for a broad spectrum of significant applications based on extensions to traditional' approaches. Multithreaded Computer Architecture is divided into four parts, reflecting four major perspectives on the topic. Part I provides the reader with basic background information, definitions, and surveys of work which have in one way or another been pivotal in defining and shaping multithreading as an architectural discipline. Part II examines key elements of multithreading, highlighting the fundamental nature of latency and synchronization. This section presents clever techniques for hiding latency and supporting large synchronization name spaces. Part III looks at three major multithreaded systems, considering issues of machine organization and compilation strategy. Part IV concludes the volume with an analysis of multithreaded architectures, showcasing methodologies and actual measurements. Multithreaded Computer Architecture: A Summary of the State of the Art is an excellent reference source and may be used as a text for advanced courses on the subject.

Fault-Tolerant Parallel and Distributed Systems (Hardcover, 1998 ed.): Dimiter R. Avresky, David R. Kaeli Fault-Tolerant Parallel and Distributed Systems (Hardcover, 1998 ed.)
Dimiter R. Avresky, David R. Kaeli
R4,559 Discovery Miles 45 590 Ships in 12 - 17 working days

The most important use of computing in the future will be in the context of the global "digital convergence" where everything becomes digital and every thing is inter-networked. The application will be dominated by storage, search, retrieval, analysis, exchange and updating of information in a wide variety of forms. Heavy demands will be placed on systems by many simultaneous re quests. And, fundamentally, all this shall be delivered at much higher levels of dependability, integrity and security. Increasingly, large parallel computing systems and networks are providing unique challenges to industry and academia in dependable computing, espe cially because of the higher failure rates intrinsic to these systems. The chal lenge in the last part of this decade is to build a systems that is both inexpensive and highly available. A machine cluster built of commodity hardware parts, with each node run ning an OS instance and a set of applications extended to be fault resilient can satisfy the new stringent high-availability requirements. The focus of this book is to present recent techniques and methods for im plementing fault-tolerant parallel and distributed computing systems. Section I, Fault-Tolerant Protocols, considers basic techniques for achieving fault-tolerance in communication protocols for distributed systems, including synchronous and asynchronous group communication, static total causal order ing protocols, and fail-aware datagram service that supports communications by time."

Performance Evaluation, Prediction and Visualization of Parallel Systems (Hardcover, 1999 ed.): Xingfu Wu Performance Evaluation, Prediction and Visualization of Parallel Systems (Hardcover, 1999 ed.)
Xingfu Wu
R4,544 Discovery Miles 45 440 Ships in 12 - 17 working days

Performance Evaluation, Prediction and Visualization in Parallel Systems presents a comprehensive and systematic discussion of theoretics, methods, techniques and tools for performance evaluation, prediction and visualization of parallel systems. Chapter 1 gives a short overview of performance degradation of parallel systems, and presents a general discussion on the importance of performance evaluation, prediction and visualization of parallel systems. Chapter 2 analyzes and defines several kinds of serial and parallel runtime, points out some of the weaknesses of parallel speedup metrics, and discusses how to improve and generalize them. Chapter 3 describes formal definitions of scalability, addresses the basic metrics affecting the scalability of parallel systems, discusses scalability of parallel systems from three aspects: parallel architecture, parallel algorithm and parallel algorithm-architecture combinations, and analyzes the relations of scalability and speedup. Chapter 4 discusses the methodology of performance measurement, describes the benchmark- oriented performance test and analysis and how to measure speedup and scalability in practice. Chapter 5 analyzes the difficulties in performance prediction, discusses application-oriented and architecture-oriented performance prediction and how to predict speedup and scalability in practice. Chapter 6 discusses performance visualization techniques and tools for parallel systems from three stages: performance data collection, performance data filtering and performance data visualization, and classifies the existing performance visualization tools. Chapter 7 describes parallel compiling-based, search-based and knowledge-based performance debugging, which assists programmers to optimize the strategy or algorithm in their parallel programs, and presents visual programming-based performance debugging to help programmers identify the location and cause of the performance problem. It also provides concrete suggestions on how to modify their parallel program to improve the performance. Chapter 8 gives an overview of current interconnection networks for parallel systems, analyzes the scalability of interconnection networks, and discusses how to measure and improve network performances. Performance Evaluation, Prediction and Visualization in Parallel Systems serves as an excellent reference for researchers, and may be used as a text for advanced courses on the topic.

Free Delivery
Pinterest Twitter Facebook Google+
You may like...
Parallel Programming in OpenMP
Rohit Chandra, Ramesh Menon, … Paperback R1,434 Discovery Miles 14 340
Computation and Storage in the Cloud…
Dong Yuan, Yun Yang, … Paperback R1,028 R869 Discovery Miles 8 690
Blockchain Applications
Arshdeep Bagha Hardcover R2,083 R1,667 Discovery Miles 16 670
Migrating Legacy Applications…
Anca Daniela Ionita, Marin Litoiu, … Hardcover R5,402 Discovery Miles 54 020
Edsger Wybe Dijkstra - His Life, Work…
Krzysztof R. Apt, Tony Hoare Hardcover R3,225 Discovery Miles 32 250
Parallel Computing: Fundamentals…
E.H. D'Hollander, G.R. Joubert, … Hardcover R6,950 Discovery Miles 69 500
Kreislauf des Lebens
Jacob Moleschott Hardcover R1,237 Discovery Miles 12 370
Programming Environments for Massively…
K.M. Decker, R. M Rehmann Hardcover R2,534 Discovery Miles 25 340
Creativity in Load-Balance Schemes for…
Alberto Garcia-Robledo, Arturo Diaz Perez, … Hardcover R4,279 Discovery Miles 42 790
Constraint Decision-Making Systems in…
Santosh Kumar Das, Nilanjan Dey Hardcover R7,388 Discovery Miles 73 880

 

Partners