0
Your cart

Your cart is empty

Browse All Departments
Price
  • R100 - R250 (4)
  • R250 - R500 (17)
  • R500+ (471)
  • -
Status
Format
Author / Contributor
Publisher

Books > Computing & IT > Computer hardware & operating systems > Computer architecture & logic design > Parallel processing

Out-of-order Parallel Discrete Event Simulation for Electronic System-level Design (Hardcover, 2015 ed.): Weiwei Chen Out-of-order Parallel Discrete Event Simulation for Electronic System-level Design (Hardcover, 2015 ed.)
Weiwei Chen
R2,867 Discovery Miles 28 670 Ships in 10 - 15 working days

This book offers readers a set of new approaches and tools a set of tools and techniques for facing challenges in parallelization with design of embedded systems. It provides an advanced parallel simulation infrastructure for efficient and effective system-level model validation and development so as to build better products in less time. Since parallel discrete event simulation (PDES) has the potential to exploit the underlying parallel computational capability in today's multi-core simulation hosts, the author begins by reviewing the parallelization of discrete event simulation, identifying problems and solutions. She then describes out-of-order parallel discrete event simulation (OoO PDES), a novel approach for efficient validation of system-level designs by aggressively exploiting the parallel capabilities of todays' multi-core PCs. This approach enables readers to design simulators that can fully exploit the parallel processing capability of the multi-core system to achieve fast speed simulation, without loss of simulation and timing accuracy. Based on this parallel simulation infrastructure, the author further describes automatic approaches that help the designer quickly to narrow down the debugging targets in faulty ESL models with parallelism.

Advances in High Performance Computing - Proceedings of the NATO Advanced Research Workshop on High Performance Computing -... Advances in High Performance Computing - Proceedings of the NATO Advanced Research Workshop on High Performance Computing - Technology and Applications, Cetraro, Italy, 24-26 June 1996 (Hardcover)
Lucio Grandinetti, Etc
R2,690 Discovery Miles 26 900 Ships in 10 - 15 working days

The book discusses some key scientific and technological developments in high performance computing, identifies significant trends, and defines desirable research objectives. It covers general concepts and emerging systems, software technology, algorithms and applications. Coverage includes hardware, software tools, networks and numerical methods, new computer architectures, and a discussion of future trends. Beyond purely scientific/engineering computing, the book extends to coverage of enterprise-wide, commercial applications, including papers on performance and scalability of database servers and Oracle DBM systems. Audience: Most papers are research level, but some are suitable for computer literate managers and technicians, making the book useful to users of commercial parallel computers.

The Sourcebook of Parallel Computing (Hardcover, New): Jack Dongarra, Ian Foster, Geoffrey C. Fox, William Gropp, Ken Kennedy,... The Sourcebook of Parallel Computing (Hardcover, New)
Jack Dongarra, Ian Foster, Geoffrey C. Fox, William Gropp, Ken Kennedy, …
R2,109 Discovery Miles 21 090 Ships in 12 - 19 working days


Parallel Computing is a compelling vision of how computation can seamlessly scale from a single processor to virtually limitless computing power. Unfortunately, the scaling of application performance has not matched peak speed, and the programming burden for these machines remains heavy. The applications must be programmed to exploit parallelism in the most efficient way possible. Today, the responsibility for achieving the vision of scalable parallelism remains in the hands of the application developer.


This book represents the collected knowledge and experience of over 60 leading parallel computing researchers. They offer students, scientists and engineers a complete sourcebook with solid coverage of parallel computing hardware, programming considerations, algorithms, software and enabling technologies, as well as several parallel application case studies. The Sourcebook of Parallel Computing offers extensive tutorials and detailed documentation of the advanced strategies produced by research over the last two decades
application case studies. The Sourcebook of Parallel Computing offers extensive tutorials and detailed documentation of the advanced strategies produced by research over the last two decades
* Provides a solid background in parallel computing technologies
* Examines the technologies available and teaches students and practitioners how to select and apply them
* Presents case studies in a range of application areas including Chemistry, Image Processing, Data Mining, Ocean Modeling and Earthquake Simulation
* Considers the future development of parallel computing technologies and the kinds of applications they will support

Distributed and Parallel Systems - From Instruction Parallelism to Cluster Computing (Hardcover, 2000 ed.): Peter Kacsuk,... Distributed and Parallel Systems - From Instruction Parallelism to Cluster Computing (Hardcover, 2000 ed.)
Peter Kacsuk, Gabriele Kotsis
R5,718 Discovery Miles 57 180 Ships in 10 - 15 working days

Distributed and Parallel Systems: From Instruction Parallelism to Cluster Computing is the proceedings of the third Austrian-Hungarian Workshop on Distributed and Parallel Systems organized jointly by the Austrian Computer Society and the MTA SZTAKI Computer and Automation Research Institute. This book contains 18 full papers and 12 short papers from 14 countries around the world, including Japan, Korea and Brazil. The paper sessions cover a broad range of research topics in the area of parallel and distributed systems, including software development environments, performance evaluation, architectures, languages, algorithms, web and cluster computing. This volume will be useful to researchers and scholars interested in all areas related to parallel and distributed computing systems.

Parallel Computing Using Optical Interconnections (Hardcover, 1998 ed.): Keqin Li, Yi Pan, Si Qing Zheng Parallel Computing Using Optical Interconnections (Hardcover, 1998 ed.)
Keqin Li, Yi Pan, Si Qing Zheng
R4,526 Discovery Miles 45 260 Ships in 10 - 15 working days

Advances in optical technologies have made it possible to implement optical interconnections in future massively parallel processing systems. Photons are non-charged particles, and do not naturally interact. Consequently, there are many desirable characteristics of optical interconnects, e.g. high speed (speed of light), increased fanout, high bandwidth, high reliability, longer interconnection lengths, low power requirements, and immunity to EMI with reduced crosstalk. Optics can utilize free-space interconnects as well as guided wave technology, neither of which has the problems of VLSI technology mentioned above. Optical interconnections can be built at various levels, providing chip-to-chip, module-to-module, board-to-board, and node-to-node communications. Massively parallel processing using optical interconnections poses new challenges; new system configurations need to be designed, scheduling and data communication schemes based on new resource metrics need to be investigated, algorithms for a wide variety of applications need to be developed under the novel computation models that optical interconnections permit, and so on. Parallel Computing Using Optical Interconnections is a collection of survey articles written by leading and active scientists in the area of parallel computing using optical interconnections. This is the first book which provides current and comprehensive coverage of the field, reflects the state of the art from high-level architecture design and algorithmic points of view, and points out directions for further research and development.

Cellular Automata - A Parallel Model (Hardcover, 1999 ed.): M. Delorme, J. Mazoyer Cellular Automata - A Parallel Model (Hardcover, 1999 ed.)
M. Delorme, J. Mazoyer
R4,572 Discovery Miles 45 720 Ships in 10 - 15 working days

Cellular automata can be viewed both as computational models and modelling systems of real processes. This volume emphasises the first aspect. In articles written by leading researchers, sophisticated massive parallel algorithms (firing squad, life, Fischer's primes recognition) are treated. Their computational power and the specific complexity classes they determine are surveyed, while some recent results in relation to chaos from a new dynamic systems point of view are also presented. Audience: This book will be of interest to specialists of theoretical computer science and the parallelism challenge.

Parallel Computer Architecture - A Hardware/Software Approach (Hardcover): David Culler, J.P. Singh, Anoop Gupta Parallel Computer Architecture - A Hardware/Software Approach (Hardcover)
David Culler, J.P. Singh, Anoop Gupta
R2,983 Discovery Miles 29 830 Ships in 12 - 19 working days

The most exciting development in parallel computer architecture is the convergence of traditionally disparate approaches on a common machine structure. This book explains the forces behind this convergence of shared-memory, message-passing, data parallel, and data-driven computing architectures. It then examines the design issues that are critical to all parallel architecture across the full range of modern design, covering data access, communication performance, coordination of cooperative work, and correct implementation of useful semantics. It not only describes the hardware and software techniques for addressing each of these issues but also explores how these techniques interact in the same system. Examining architecture from an application-driven perspective, it provides comprehensive discussions of parallel programming for high performance and of workload-driven evaluation, based on understanding hardware-software interactions.
* synthesizes a decade of research and development for practicing engineers, graduate students, and researchers in parallel computer architecture, system software, and applications development
* presents in-depth application case studies from computer graphics, computational science and engineering, and data mining to demonstrate sound quantitative evaluation of design trade-offs
* describes the process of programming for performance, including both the architecture-independent and architecture-dependent aspects, with examples and case-studies
* illustrates bus-based and network-based parallel systems with case studies of more than a dozen important commercial designs

Parallel Numerical Computation with Applications (Hardcover, 1999 ed.): Laurence Tianruo Yang Parallel Numerical Computation with Applications (Hardcover, 1999 ed.)
Laurence Tianruo Yang
R4,489 Discovery Miles 44 890 Ships in 10 - 15 working days

Parallel Numerical Computations with Applications contains selected edited papers presented at the 1998 Frontiers of Parallel Numerical Computations and Applications Workshop, along with invited papers from leading researchers around the world. These papers cover a broad spectrum of topics on parallel numerical computation with applications; such as advanced parallel numerical and computational optimization methods, novel parallel computing techniques, numerical fluid mechanics, and other applications related to material sciences, signal and image processing, semiconductor technology, and electronic circuits and systems design. This state-of-the-art volume will be an up-to-date resource for researchers in the areas of parallel and distributed computing.

Development and Deployment of Multiplayer Online Games, Vol. II - DIY, (Re)Actors, Client Arch., Unity/UE4/ Lumberyard/Urho3D... Development and Deployment of Multiplayer Online Games, Vol. II - DIY, (Re)Actors, Client Arch., Unity/UE4/ Lumberyard/Urho3D (Hardcover)
'No Bugs' Hare; Contributions by Sergey Ignatchenko
R1,061 R923 Discovery Miles 9 230 Save R138 (13%) Ships in 10 - 15 working days
Parallel and Distributed Information Systems (Hardcover, Reprinted From Parallel And Distributed Databases,): Jeffrey F.... Parallel and Distributed Information Systems (Hardcover, Reprinted From Parallel And Distributed Databases,)
Jeffrey F. Naughton, Gerhard Weikum
R2,934 Discovery Miles 29 340 Ships in 10 - 15 working days

Parallel and Distributed Information Systems brings together in one place important contributions and up-to-date research results in this fast moving area. Parallel and Distributed Information Systems serves as an excellent reference, providing insight into some of the most challenging research issues in the field.

Concurrent Scientific Computing (Hardcover, 1994 ed.): Eric F. Van De Velde Concurrent Scientific Computing (Hardcover, 1994 ed.)
Eric F. Van De Velde
R2,117 Discovery Miles 21 170 Ships in 12 - 19 working days

Mathematics is playing an ever more important role in the physical and biological sciences, provoking a blurring of boundaries between scientific dis ciplines and a resurgence of interest in the modern as well as the classical techniques of applied mathematics. This renewal of interest, both in research and teaching, has led to the establishment of the series: Texts in Applied Mathe matics (TAM). The development of new courses is a natural consequence of a high level of excitement on the research frontier as newer techniques, such as numerical and symbolic computer systems, dynamical systems, and chaos, mix with and reinforce the traditional methods of applied mathematics. Thus, the purpose of this textbook series is to meet the current and future needs of these advances and encourage the teaching of new courses. TAM will publish textbooks suitable for use in advanced undergraduate and beginning graduate courses, and will complement the Applied Mathematical Sciences (AMS) series, which will focus on advanced textbooks and research level monographs. Preface A successful concurrent numerical simulation requires physics and math ematics to develop and analyze the model, numerical analysis to develop solution methods, and computer science to develop a concurrent implemen tation. No single course can or should cover all these disciplines. Instead, this course on concurrent scientific computing focuses on a topic that is not covered or is insufficiently covered by other disciplines: the algorith mic structure of numerical methods."

Parallel Computing on Distributed Memory Multiprocessors (Hardcover, 1993 ed.): Fusun Oezguner, Fikret Ercal Parallel Computing on Distributed Memory Multiprocessors (Hardcover, 1993 ed.)
Fusun Oezguner, Fikret Ercal
R3,073 Discovery Miles 30 730 Ships in 10 - 15 working days

Advances in microelectronic technology have made massively parallel computing a reality and triggered an outburst of research activity in parallel processing architectures and algorithms. Distributed memory multiprocessors - parallel computers that consist of microprocessors connected in a regular topology - are increasingly being used to solve large problems in many application areas. In order to use these computers for a specific application, existing algorithms need to be restructured for the architecture and new algorithms developed. The performance of a computation on a distributed memory multiprocessor is affected by the node and communication architecture, the interconnection network topology, the I/O subsystem, and the parallel algorithm and communication protocols. Each of these parametersis a complex problem, and solutions require an understanding of the interactions among them. This book is based on the papers presented at the NATO Advanced Study Institute held at Bilkent University, Turkey, in July 1991. The book is organized in five parts: Parallel computing structures and communication, Parallel numerical algorithms, Parallel programming, Fault tolerance, and Applications and algorithms.

Algorithms for Parallel Processing (Hardcover, 1999 ed.): Michael T. Heath, Abhiram Ranade, Robert S. Schreiber Algorithms for Parallel Processing (Hardcover, 1999 ed.)
Michael T. Heath, Abhiram Ranade, Robert S. Schreiber
R3,097 Discovery Miles 30 970 Ships in 10 - 15 working days

This IMA Volume in Mathematics and its Applications ALGORITHMS FOR PARALLEL PROCESSING is based on the proceedings of a workshop that was an integral part of the 1996-97 IMA program on "MATHEMATICS IN HIGH-PERFORMANCE COMPUTING. " The workshop brought together algorithm developers from theory, combinatorics, and scientific computing. The topics ranged over models, linear algebra, sorting, randomization, and graph algorithms and their analysis. We thank Michael T. Heath of University of lllinois at Urbana (Com puter Science), Abhiram Ranade of the Indian Institute of Technology (Computer Science and Engineering), and Robert S. Schreiber of Hewlett Packard Laboratories for their excellent work in organizing the workshop and editing the proceedings. We also take this opportunity to thank the National Science Founda tion (NSF) and the Army Research Office (ARO), whose financial support made the workshop possible. A vner Friedman Robert Gulliver v PREFACE The Workshop on Algorithms for Parallel Processing was held at the IMA September 16 - 20, 1996; it was the first workshop of the IMA year dedicated to the mathematics of high performance computing. The work shop organizers were Abhiram Ranade of The Indian Institute of Tech nology, Bombay, Michael Heath of the University of Illinois, and Robert Schreiber of Hewlett Packard Laboratories. Our idea was to bring together researchers who do innovative, exciting, parallel algorithms research on a wide range of topics, and by sharing insights, problems, tools, and methods to learn something of value from one another."

Parallel Computing in Optimization (Hardcover, 1997 ed.): A. Migdalas, Panos M. Pardalos, Sverre Storoy Parallel Computing in Optimization (Hardcover, 1997 ed.)
A. Migdalas, Panos M. Pardalos, Sverre Storoy
R8,638 Discovery Miles 86 380 Ships in 10 - 15 working days

During the last three decades, breakthroughs in computer technology have made a tremendous impact on optimization. In particular, parallel computing has made it possible to solve larger and computationally more difficult problems. The book covers recent developments in novel programming and algorithmic aspects of parallel computing as well as technical advances in parallel optimization. Each contribution is essentially expository in nature, but of scholarly treatment. In addition, each chapter includes a collection of carefully selected problems. The first two chapters discuss theoretical models for parallel algorithm design and their complexity. The next chapter gives the perspective of the programmer practicing parallel algorithm development on real world platforms. Solving systems of linear equations efficiently is of great importance not only because they arise in many scientific and engineering applications but also because algorithms for solving many optimization problems need to call system solvers and subroutines (chapters four and five). Chapters six through thirteen are dedicated to optimization problems and methods. They include parallel algorithms for network problems, parallel branch and bound techniques, parallel heuristics for discrete and continuous problems, decomposition methods, parallel algorithms for variational inequality problems, parallel algorithms for stochastic programming, and neural networks. Audience: Parallel Computing in Optimization is addressed not only to researchers of mathematical programming, but to all scientists in various disciplines who use optimization methods in parallel and multiprocessing environments to model and solve problems.

Software-Implemented Hardware Fault Tolerance (Hardcover): Olga Goloubeva, Maurizio Rebaudengo, Matteo Sonza Reorda, Massimo... Software-Implemented Hardware Fault Tolerance (Hardcover)
Olga Goloubeva, Maurizio Rebaudengo, Matteo Sonza Reorda, Massimo Violante
R4,485 Discovery Miles 44 850 Ships in 10 - 15 working days

This book presents the theory behind software-implemented hardware fault tolerance, as well as the practical aspects needed to put it to work on real examples. By evaluating accurately the advantages and disadvantages of the already available approaches, the book provides a guide to developers willing to adopt software-implemented hardware fault tolerance in their applications. Moreover, the book identifies open issues for researchers willing to improve the already available techniques.

Hierarchical Scheduling in Parallel and Cluster Systems (Hardcover, 2003 ed.): Sivarama Dandamudi Hierarchical Scheduling in Parallel and Cluster Systems (Hardcover, 2003 ed.)
Sivarama Dandamudi
R4,507 Discovery Miles 45 070 Ships in 10 - 15 working days

Multiple processor systems are an important class of parallel systems. Over the years, several architectures have been proposed to build such systems to satisfy the requirements of high performance computing. These architectures span a wide variety of system types. At the low end of the spectrum, we can build a small, shared-memory parallel system with tens of processors. These systems typically use a bus to interconnect the processors and memory. Such systems, for example, are becoming commonplace in high-performance graph ics workstations. These systems are called uniform memory access (UMA) multiprocessors because they provide uniform access of memory to all pro cessors. These systems provide a single address space, which is preferred by programmers. This architecture, however, cannot be extended even to medium systems with hundreds of processors due to bus bandwidth limitations. To scale systems to medium range i. e., to hundreds of processors, non-bus interconnection networks have been proposed. These systems, for example, use a multistage dynamic interconnection network. Such systems also provide global, shared memory like the UMA systems. However, they introduce local and remote memories, which lead to non-uniform memory access (NUMA) architecture. Distributed-memory architecture is used for systems with thousands of pro cessors. These systems differ from the shared-memory architectures in that there is no globally accessible shared memory. Instead, they use message pass ing to facilitate communication among the processors. As a result, they do not provide single address space."

Introduction to Parallel Processing - Algorithms and Architectures (Hardcover, 1999 ed.): Behrooz Parhami Introduction to Parallel Processing - Algorithms and Architectures (Hardcover, 1999 ed.)
Behrooz Parhami
R6,041 Discovery Miles 60 410 Ships in 10 - 15 working days

THE CONTEXT OF PARALLEL PROCESSING The field of digital computer architecture has grown explosively in the past two decades. Through a steady stream of experimental research, tool-building efforts, and theoretical studies, the design of an instruction-set architecture, once considered an art, has been transformed into one of the most quantitative branches of computer technology. At the same time, better understanding of various forms of concurrency, from standard pipelining to massive parallelism, and invention of architectural structures to support a reasonably efficient and user-friendly programming model for such systems, has allowed hardware performance to continue its exponential growth. This trend is expected to continue in the near future. This explosive growth, linked with the expectation that performance will continue its exponential rise with each new generation of hardware and that (in stark contrast to software) computer hardware will function correctly as soon as it comes off the assembly line, has its down side. It has led to unprecedented hardware complexity and almost intolerable dev- opment costs. The challenge facing current and future computer designers is to institute simplicity where we now have complexity; to use fundamental theories being developed in this area to gain performance and ease-of-use benefits from simpler circuits; to understand the interplay between technological capabilities and limitations, on the one hand, and design decisions based on user and application requirements on the other.

Tools and Environments for Parallel and Distributed Systems (Hardcover, 1996 ed.): Amr Zaky, Ted Lewis Tools and Environments for Parallel and Distributed Systems (Hardcover, 1996 ed.)
Amr Zaky, Ted Lewis
R4,530 Discovery Miles 45 300 Ships in 10 - 15 working days

Developing correct and efficient software is far more complex for parallel and distributed systems than it is for sequential processors. Some of the reasons for this added complexity are: the lack of a universally acceptable parallel and distributed programming paradigm, the criticality of achieving high performance, and the difficulty of writing correct parallel and distributed programs. These factors collectively influence the current status of parallel and distributed software development tools efforts. Tools and Environments for Parallel and Distributed Systems addresses the above issues by describing working tools and environments, and gives a solid overview of some of the fundamental research being done worldwide. Topics covered in this collection are: mainstream program development tools, performance prediction tools and studies; debugging tools and research; and nontraditional tools. Audience: Suitable as a secondary text for graduate level courses in software engineering and parallel and distributed systems, and as a reference for researchers and practitioners in industry.

Multithreaded Computer Architecture: A Summary of the State of the ART (Hardcover, 1994 ed.): Robert A. Iannucci, Guang R. Gao,... Multithreaded Computer Architecture: A Summary of the State of the ART (Hardcover, 1994 ed.)
Robert A. Iannucci, Guang R. Gao, Robert H. Halstead Jr, Burton Smith
R5,820 Discovery Miles 58 200 Ships in 10 - 15 working days

Multithreaded computer architecture has emerged as one of the most promising and exciting avenues for the exploitation of parallelism. This new field represents the confluence of several independent research directions which have united over a common set of issues and techniques. Multithreading draws on recent advances in dataflow, RISC, compiling for fine-grained parallel execution, and dynamic resource management. It offers the hope of dramatic performance increases through parallel execution for a broad spectrum of significant applications based on extensions to traditional' approaches. Multithreaded Computer Architecture is divided into four parts, reflecting four major perspectives on the topic. Part I provides the reader with basic background information, definitions, and surveys of work which have in one way or another been pivotal in defining and shaping multithreading as an architectural discipline. Part II examines key elements of multithreading, highlighting the fundamental nature of latency and synchronization. This section presents clever techniques for hiding latency and supporting large synchronization name spaces. Part III looks at three major multithreaded systems, considering issues of machine organization and compilation strategy. Part IV concludes the volume with an analysis of multithreaded architectures, showcasing methodologies and actual measurements. Multithreaded Computer Architecture: A Summary of the State of the Art is an excellent reference source and may be used as a text for advanced courses on the subject.

Fault-Tolerant Parallel and Distributed Systems (Hardcover, 1998 ed.): Dimiter R. Avresky, David R. Kaeli Fault-Tolerant Parallel and Distributed Systems (Hardcover, 1998 ed.)
Dimiter R. Avresky, David R. Kaeli
R4,589 Discovery Miles 45 890 Ships in 10 - 15 working days

The most important use of computing in the future will be in the context of the global "digital convergence" where everything becomes digital and every thing is inter-networked. The application will be dominated by storage, search, retrieval, analysis, exchange and updating of information in a wide variety of forms. Heavy demands will be placed on systems by many simultaneous re quests. And, fundamentally, all this shall be delivered at much higher levels of dependability, integrity and security. Increasingly, large parallel computing systems and networks are providing unique challenges to industry and academia in dependable computing, espe cially because of the higher failure rates intrinsic to these systems. The chal lenge in the last part of this decade is to build a systems that is both inexpensive and highly available. A machine cluster built of commodity hardware parts, with each node run ning an OS instance and a set of applications extended to be fault resilient can satisfy the new stringent high-availability requirements. The focus of this book is to present recent techniques and methods for im plementing fault-tolerant parallel and distributed computing systems. Section I, Fault-Tolerant Protocols, considers basic techniques for achieving fault-tolerance in communication protocols for distributed systems, including synchronous and asynchronous group communication, static total causal order ing protocols, and fail-aware datagram service that supports communications by time."

Performance Evaluation, Prediction and Visualization of Parallel Systems (Hardcover, 1999 ed.): Xingfu Wu Performance Evaluation, Prediction and Visualization of Parallel Systems (Hardcover, 1999 ed.)
Xingfu Wu
R4,548 Discovery Miles 45 480 Ships in 10 - 15 working days

Performance Evaluation, Prediction and Visualization in Parallel Systems presents a comprehensive and systematic discussion of theoretics, methods, techniques and tools for performance evaluation, prediction and visualization of parallel systems. Chapter 1 gives a short overview of performance degradation of parallel systems, and presents a general discussion on the importance of performance evaluation, prediction and visualization of parallel systems. Chapter 2 analyzes and defines several kinds of serial and parallel runtime, points out some of the weaknesses of parallel speedup metrics, and discusses how to improve and generalize them. Chapter 3 describes formal definitions of scalability, addresses the basic metrics affecting the scalability of parallel systems, discusses scalability of parallel systems from three aspects: parallel architecture, parallel algorithm and parallel algorithm-architecture combinations, and analyzes the relations of scalability and speedup. Chapter 4 discusses the methodology of performance measurement, describes the benchmark- oriented performance test and analysis and how to measure speedup and scalability in practice. Chapter 5 analyzes the difficulties in performance prediction, discusses application-oriented and architecture-oriented performance prediction and how to predict speedup and scalability in practice. Chapter 6 discusses performance visualization techniques and tools for parallel systems from three stages: performance data collection, performance data filtering and performance data visualization, and classifies the existing performance visualization tools. Chapter 7 describes parallel compiling-based, search-based and knowledge-based performance debugging, which assists programmers to optimize the strategy or algorithm in their parallel programs, and presents visual programming-based performance debugging to help programmers identify the location and cause of the performance problem. It also provides concrete suggestions on how to modify their parallel program to improve the performance. Chapter 8 gives an overview of current interconnection networks for parallel systems, analyzes the scalability of interconnection networks, and discusses how to measure and improve network performances. Performance Evaluation, Prediction and Visualization in Parallel Systems serves as an excellent reference for researchers, and may be used as a text for advanced courses on the topic.

Multiprocessing - Trade-Offs in Computation and Communication (Hardcover, 1993 ed.): Vijay K. Naik Multiprocessing - Trade-Offs in Computation and Communication (Hardcover, 1993 ed.)
Vijay K. Naik
R3,000 Discovery Miles 30 000 Ships in 10 - 15 working days

Multiprocessing: Trade-Offs in Computation and Communication presents an in-depth analysis of several commonly observed regular and irregular computations for multiprocessor systems. This book includes techniques which enable researchers and application developers to quantitatively determine the effects of algorithm data dependencies on execution time, on communication requirements, on processor utilization and on the speedups possible. Starting with simple, two-dimensional, diamond-shaped directed acyclic graphs, the analysis is extended to more complex and higher dimensional directed acyclic graphs. The analysis allows for the quantification of the computation and communication costs and their interdependencies. The practical significance of these results on the performance of various data distribution schemes is clearly explained. Using these results, the performance of the parallel computations are formulated in an architecture independent fashion. These formulations allow for the parameterization of the architecture specitific entities such as the computation and communication rates. This type of parameterized performance analysis can be used at compile time or at run-time so as to achieve the most optimal distribution of the computations. The material in Multiprocessing: Trade-Offs in Computation and Communication connects theory with practice, so that the inherent performance limitations in many computations can be understood, and practical methods can be devised that would assist in the development of software for scalable high performance systems.

Communication Complexity and Parallel Computing (Hardcover, 1997 ed.): Juraj Hromkovic Communication Complexity and Parallel Computing (Hardcover, 1997 ed.)
Juraj Hromkovic
R1,725 Discovery Miles 17 250 Ships in 10 - 15 working days

The communication complexity of two-party protocols is an only 15 years old complexity measure, but it is already considered to be one of the fundamen tal complexity measures of recent complexity theory. Similarly to Kolmogorov complexity in the theory of sequential computations, communication complex ity is used as a method for the study of the complexity of concrete computing problems in parallel information processing. Especially, it is applied to prove lower bounds that say what computer resources (time, hardware, memory size) are necessary to compute the given task. Besides the estimation of the compu tational difficulty of computing problems the proved lower bounds are useful for proving the optimality of algorithms that are already designed. In some cases the knowledge about the communication complexity of a given problem may be even helpful in searching for efficient algorithms to this problem. The study of communication complexity becomes a well-defined indepen dent area of complexity theory. In addition to a strong relation to several funda mental complexity measures (and so to several fundamental problems of com plexity theory) communication complexity has contributed to the study and to the understanding of the nature of determinism, nondeterminism, and random ness in algorithmics. There already exists a non-trivial mathematical machinery to handle the communication complexity of concrete computing problems, which gives a hope that the approach based on communication complexity will be in strumental in the study of several central open problems of recent complexity theory."

Nonlinear Assignment Problems - Algorithms and Applications (Hardcover, 2001 ed.): Panos M. Pardalos, L.S. Pitsoulis Nonlinear Assignment Problems - Algorithms and Applications (Hardcover, 2001 ed.)
Panos M. Pardalos, L.S. Pitsoulis
R4,390 Discovery Miles 43 900 Ships in 10 - 15 working days

Nonlinear Assignment Problems (NAPs) are natural extensions of the classic Linear Assignment Problem, and despite the efforts of many researchers over the past three decades, they still remain some of the hardest combinatorial optimization problems to solve exactly. The purpose of this book is to provide in a single volume, major algorithmic aspects and applications of NAPs as contributed by leading international experts. The chapters included in this book are concerned with major applications and the latest algorithmic solution approaches for NAPs. Approximation algorithms, polyhedral methods, semidefinite programming approaches and heuristic procedures for NAPs are included, while applications of this problem class in the areas of multiple-target tracking in the context of military surveillance systems, of experimental high energy physics, and of parallel processing are presented. Audience: Researchers and graduate students in the areas of combinatorial optimization, mathematical programming, operations research, physics, and computer science.

A Parallel Algorithm Synthesis Procedure for High-Performance Computer Architectures (Hardcover, 2003 ed.): Ian N. Dunn, Gerard... A Parallel Algorithm Synthesis Procedure for High-Performance Computer Architectures (Hardcover, 2003 ed.)
Ian N. Dunn, Gerard G.L. Meyer
R2,941 Discovery Miles 29 410 Ships in 10 - 15 working days

Despite five decades of research, parallel computing remains an exotic, frontier technology on the fringes of mainstream computing. Its much-heralded triumph over sequential computing has yet to materialize. This is in spite of the fact that the processing needs of many signal processing applications continue to eclipse the capabilities of sequential computing. The culprit is largely the software development environment. Fundamental shortcomings in the development environment of many parallel computer architectures thwart the adoption of parallel computing. Foremost, parallel computing has no unifying model to accurately predict the execution time of algorithms on parallel architectures. Cost and scarce programming resources prohibit deploying multiple algorithms and partitioning strategies in an attempt to find the fastest solution. As a consequence, algorithm design is largely an intuitive art form dominated by practitioners who specialize in a particular computer architecture. This, coupled with the fact that parallel computer architectures rarely last more than a couple of years, makes for a complex and challenging design environment.

To navigate this environment, algorithm designers need a road map, a detailed procedure they can use to efficiently develop high performance, portable parallel algorithms. The focus of this book is to draw such a road map. The Parallel Algorithm Synthesis Procedure can be used to design reusable building blocks of adaptable, scalable software modules from which high performance signal processing applications can be constructed. The hallmark of the procedure is a semi-systematic process for introducing parameters to control the partitioning andscheduling of computation and communication. This facilitates the tailoring of software modules to exploit different configurations of multiple processors, multiple floating-point units, and hierarchical memories. To showcase the efficacy of this procedure, the book presents three case studies requiring various degrees of optimization for parallel execution.

This book can be used as a reference for algorithm designers or as a text for an advanced course on parallel programming.

Free Delivery
Pinterest Twitter Facebook Google+
You may like...
Probabilistic Nodes Combination (PNC…
Dariusz Jacek Jakobczak Hardcover R4,776 Discovery Miles 47 760
Java - The ultimate beginners guide to…
Mark Reed Hardcover R613 R556 Discovery Miles 5 560
Python Programming For Beginners In 2021…
James Tudor Hardcover R724 Discovery Miles 7 240
Java Cookbook - Problems and Solutions…
Ian F. Darwin Paperback R1,965 R1,608 Discovery Miles 16 080
Understanding Object-Oriented…
Timothy Budd Paperback R2,243 R2,084 Discovery Miles 20 840
Formal Methods for Open Object-Based…
Paolo Ciancarini, Alessandro Fantechi, … Hardcover R5,838 Discovery Miles 58 380
Object-Oriented Game Development
Julian Gold Paperback R2,247 R1,778 Discovery Miles 17 780
Object-Oriented Technology
Suad Alagi c Hardcover R1,982 Discovery Miles 19 820
Enterprise Modeling and Computing with…
Hardcover R2,697 Discovery Miles 26 970
Head First Java, 3rd Edition - A…
Kathy Sierra, Bert Bates, … Paperback R1,622 R1,431 Discovery Miles 14 310

 

Partners