![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer hardware & operating systems > Supercomputers
From the Foreword: "The authors of the chapters in this book are the pioneers who will explore the exascale frontier. The path forward will not be easy... These authors, along with their colleagues who will produce these powerful computer systems will, with dedication and determination, overcome the scalability problem, discover the new algorithms needed to achieve exascale performance for the broad range of applications that they represent, and create the new tools needed to support the development of scalable and portable science and engineering applications. Although the focus is on exascale computers, the benefits will permeate all of science and engineering because the technologies developed for the exascale computers of tomorrow will also power the petascale servers and terascale workstations of tomorrow. These affordable computing capabilities will empower scientists and engineers everywhere." - Thom H. Dunning, Jr., Pacific Northwest National Laboratory and University of Washington, Seattle, Washington, USA "This comprehensive summary of applications targeting Exascale at the three DoE labs is a must read." - Rio Yokota, Tokyo Institute of Technology, Tokyo, Japan "Numerical simulation is now a need in many fields of science, technology, and industry. The complexity of the simulated systems coupled with the massive use of data makes HPC essential to move towards predictive simulations. Advances in computer architecture have so far permitted scientific advances, but at the cost of continually adapting algorithms and applications. The next technological breakthroughs force us to rethink the applications by taking energy consumption into account. These profound modifications require not only anticipation and sharing but also a paradigm shift in application design to ensure the sustainability of developments by guaranteeing a certain independence of the applications to the profound modifications of the architectures: it is the passage from optimal performance to the portability of performance. It is the challenge of this book to demonstrate by example the approach that one can adopt for the development of applications offering performance portability in spite of the profound changes of the computing architectures." - Christophe Calvin, CEA, Fundamental Research Division, Saclay, France "Three editors, one from each of the High Performance Computer Centers at Lawrence Berkeley, Argonne, and Oak Ridge National Laboratories, have compiled a very useful set of chapters aimed at describing software developments for the next generation exa-scale computers. Such a book is needed for scientists and engineers to see where the field is going and how they will be able to exploit such architectures for their own work. The book will also benefit students as it provides insights into how to develop software for such computer architectures. Overall, this book fills an important need in showing how to design and implement algorithms for exa-scale architectures which are heterogeneous and have unique memory systems. The book discusses issues with developing user codes for these architectures and how to address these issues including actual coding examples.' - Dr. David A. Dixon, Robert Ramsay Chair, The University of Alabama, Tuscaloosa, Alabama, USA
This multi-contributed handbook focuses on the latest workings of IoT (internet of Things) and Big Data. As the resources are limited, it's the endeavor of the authors to support and bring the information into one resource. The book is divided into 4 sections that covers IoT and technologies, the future of Big Data, algorithms, and case studies showing IoT and Big Data in various fields such as health care, manufacturing and automation. Features Focuses on the latest workings of IoT and Big Data Discusses the emerging role of technologies and the fast-growing market of Big Data Covers the movement toward automation with hardware, software, and sensors, and trying to save on energy resources Offers the latest technology on IoT Presents the future horizons on Big Data
The Future of Numerical Computing Written by one of the foremost experts in high-performance computing and the inventor of Gustafson's Law, The End of Error: Unum Computing explains a new approach to computer arithmetic: the universal number (unum). The unum encompasses all IEEE floating-point formats as well as fixed-point and exact integer arithmetic. This new number type obtains more accurate answers than floating-point arithmetic yet uses fewer bits in many cases, saving memory, bandwidth, energy, and power. A Complete Revamp of Computer Arithmetic from the Ground Up Richly illustrated in color, this groundbreaking book represents a fundamental change in how to perform calculations automatically. It illustrates how this novel approach can solve problems that have vexed engineers and scientists for decades, including problems that have been historically limited to serial processing. Suitable for Anyone Using Computers for Calculations The book is accessible to anyone who uses computers for technical calculations, with much of the book only requiring high school math. The author makes the mathematics interesting through numerous analogies. He clearly defines jargon and uses color-coded boxes for mathematical formulas, computer code, important descriptions, and exercises.
Introduction to Modeling and Simulation with MATLAB and Python is intended for students and professionals in science, social science, and engineering that wish to learn the principles of computer modeling, as well as basic programming skills. The book content focuses on meeting a set of basic modeling and simulation competencies that were developed as part of several National Science Foundation grants. Even though computer science students are much more expert programmers, they are not often given the opportunity to see how those skills are being applied to solve complex science and engineering problems and may also not be aware of the libraries used by scientists to create those models. The book interleaves chapters on modeling concepts and related exercises with programming concepts and exercises. The authors start with an introduction to modeling and its importance to current practices in the sciences and engineering. They introduce each of the programming environments and the syntax used to represent variables and compute mathematical equations and functions. As students gain more programming expertise, the authors return to modeling concepts, providing starting code for a variety of exercises where students add additional code to solve the problem and provide an analysis of the outcomes. In this way, the book builds both modeling and programming expertise with a "just-in-time" approach so that by the end of the book, students can take on relatively simple modeling example on their own. Each chapter is supplemented with references to additional reading, tutorials, and exercises that guide students to additional help and allows them to practice both their programming and analytical modeling skills. In addition, each of the programming related chapters is divided into two parts - one for MATLAB and one for Python. In these chapters, the authors also refer to additional online tutorials that students can use if they are having difficulty with any of the topics. The book culminates with a set of final project exercise suggestions that incorporate both the modeling and programming skills provided in the rest of the volume. Those projects could be undertaken by individuals or small groups of students. The companion website at http://www.intromodeling.com provides updates to instructions when there are substantial changes in software versions, as well as electronic copies of exercises and the related code. The website also offers a space where people can suggest additional projects they are willing to share as well as comments on the existing projects and exercises throughout the book. Solutions and lecture notes will also be available for qualifying instructors.
Combinatorial Scientific Computing explores the latest research on creating algorithms and software tools to solve key combinatorial problems on large-scale high-performance computing architectures. It includes contributions from international researchers who are pioneers in designing software and applications for high-performance computing systems. The book offers a state-of-the-art overview of the latest research, tool development, and applications. It focuses on load balancing and parallelization on high-performance computers, large-scale optimization, algorithmic differentiation of numerical simulation code, sparse matrix software tools, and combinatorial challenges and applications in large-scale social networks. The authors unify these seemingly disparate areas through a common set of abstractions and algorithms based on combinatorics, graphs, and hypergraphs. Combinatorial algorithms have long played a crucial enabling role in scientific and engineering computations and their importance continues to grow with the demands of new applications and advanced architectures. By addressing current challenges in the field, this volume sets the stage for the accelerated development and deployment of fundamental enabling technologies in high-performance scientific computing.
Exploring how concurrent programming can be assisted by language-level techniques, Introduction to Concurrency in Programming Languages presents high-level language techniques for dealing with concurrency in a general context. It provides an understanding of programming languages that offer concurrency features as part of the language definition. The book supplies a conceptual framework for different aspects of parallel algorithm design and implementation. It first addresses the limitations of traditional programming techniques and models when dealing with concurrency. The book then explores the current state of the art in concurrent programming and describes high-level language constructs for concurrency. It also discusses the historical evolution of hardware, corresponding high-level techniques that were developed, and the connection to modern systems, such as multicore and manycore processors. The remainder of the text focuses on common high-level programming techniques and their application to a range of algorithms. The authors offer case studies on genetic algorithms, fractal generation, cellular automata, game logic for solving Sudoku puzzles, pipelined algorithms, and more. Illustrating the effect of concurrency on programs written in familiar languages, this text focuses on novel language abstractions that truly bring concurrency into the language and aid analysis and compilation tools in generating efficient, correct programs. It also explains the complexity involved in taking advantage of concurrency with regard to program correctness and performance.
As more and more data is generated at a faster-than-ever rate, processing large volumes of data is becoming a challenge for data analysis software. Addressing performance issues, Cloud Computing: Data-Intensive Computing and Scheduling explores the evolution of classical techniques and describes completely new methods and innovative algorithms. The book delineates many concepts, models, methods, algorithms, and software used in cloud computing. After a general introduction to the field, the text covers resource management, including scheduling algorithms for real-time tasks and practical algorithms for user bidding and auctioneer pricing. It next explains approaches to data analytical query processing, including pre-computing, data indexing, and data partitioning. Applications of MapReduce, a new parallel programming model, are then presented. The authors also discuss how to optimize multiple group-by query processing and introduce a MapReduce real-time scheduling algorithm. A useful reference for studying and using MapReduce and cloud computing platforms, this book presents various technologies that demonstrate how cloud computing can meet business requirements and serve as the infrastructure of multidimensional data analysis applications.
A bold, visionary, and mind-bending exploration of how the geometry of chaos can explain our uncertain world - from weather and pandemics to quantum physics and free will Covering a breathtaking range of topics - from climate change to the foundations of quantum physics, from economic modelling to conflict prediction, from free will to consciousness and spirituality - The Primacy of Doubt takes us on a unique journey through the science of uncertainty. A key theme that unifies these seemingly unconnected topics is the geometry of chaos: the beautiful and profound fractal structures that lie at the heart of much of modern mathematics. Royal Society Research Professor Tim Palmer shows us how the geometry of chaos not only provides the means to predict the world around us, it suggests new insights into some of the most astonishing aspects of our universe and ourselves. This important and timely book helps the reader makes sense of uncertainty in a rapidly changing world.
This practical book presents fundamental concepts and issues in computer modeling and simulation (M&S) in a simple and practical way for engineers, scientists, and managers who wish to apply simulation successfully to their real-world problems. It offers a concise approach to the coverage of generic (tool-independent) M&S concepts and enables engineering practitioners to easily learn, evaluate, and apply various available simulation concepts. Worked out examples are included to illustrate the concepts and an example modeling application is continued throughout the chapters to demonstrate the techniques. The book discusses modeling purposes, scoping a model, levels of modeling abstraction, the benefits and cost of including randomness, types of simulation, and statistical techniques. It also includes a chapter on modeling and simulation projects and how to conduct them for customer and engineer benefit and covers the stages of a modeling and simulation study, including process and system investigation, data collection, modeling scoping and production, model verification and validation, experimentation, and analysis of results.
This practical book presents fundamental concepts and issues in computer modeling and simulation (M&S) in a simple and practical way for engineers, scientists, and managers who wish to apply simulation successfully to their real-world problems. It offers a concise approach to the coverage of generic (tool-independent) M&S concepts and enables engineering practitioners to easily learn, evaluate, and apply various available simulation concepts. Worked out examples are included to illustrate the concepts and an example modeling application is continued throughout the chapters to demonstrate the techniques. The book discusses modeling purposes, scoping a model, levels of modeling abstraction, the benefits and cost of including randomness, types of simulation, and statistical techniques. It also includes a chapter on modeling and simulation projects and how to conduct them for customer and engineer benefit and covers the stages of a modeling and simulation study, including process and system investigation, data collection, modeling scoping and production, model verification and validation, experimentation, and analysis of results.
High-Performance Computing for Big Data: Methodologies and Applications explores emerging high-performance architectures for data-intensive applications, novel efficient analytical strategies to boost data processing, and cutting-edge applications in diverse fields, such as machine learning, life science, neural networks, and neuromorphic engineering. The book is organized into two main sections. The first section covers Big Data architectures, including cloud computing systems, and heterogeneous accelerators. It also covers emerging 3D IC design principles for memory architectures and devices. The second section of the book illustrates emerging and practical applications of Big Data across several domains, including bioinformatics, deep learning, and neuromorphic engineering. Features Covers a wide range of Big Data architectures, including distributed systems like Hadoop/Spark Includes accelerator-based approaches for big data applications such as GPU-based acceleration techniques, and hardware acceleration such as FPGA/CGRA/ASICs Presents emerging memory architectures and devices such as NVM, STT- RAM, 3D IC design principles Describes advanced algorithms for different big data application domains Illustrates novel analytics techniques for Big Data applications, scheduling, mapping, and partitioning methodologies Featuring contributions from leading experts, this book presents state-of-the-art research on the methodologies and applications of high-performance computing for big data applications. About the Editor Dr. Chao Wang is an Associate Professor in the School of Computer Science at the University of Science and Technology of China. He is the Associate Editor of ACM Transactions on Design Automations for Electronics Systems (TODAES), Applied Soft Computing, Microprocessors and Microsystems, IET Computers & Digital Techniques, and International Journal of Electronics. Dr. Chao Wang was the recipient of Youth Innovation Promotion Association, CAS, ACM China Rising Star Honorable Mention (2016), and best IP nomination of DATE 2015. He is now on the CCF Technical Committee on Computer Architecture, CCF Task Force on Formal Methods. He is a Senior Member of IEEE, Senior Member of CCF, and a Senior Member of ACM.
This book constitutes the refereed conference proceedings of the workshops held at the 37th International ISC High Performance 2022 Conference, in Hamburg, Germany, in June 2, 2022. The 27 full papers were included in this book were carefully reviewed and selected from 43 submissions. ISC High Performance 2022 presents the following workshops: Compiler-assisted Correctness Checking and Performance Optimization for HPC HPC on Heterogeneous Hardware (H3) Malleability Techniques Applications in High Performance Computing Fifth Workshop on Interactive High Performance Computing 3rd ISC HPC International Workshop on Monitoring & Operational Data Analytics 6th International Workshop on In Situ Visualization 17th Workshop on Virtualization in High Performance Cloud Computing
Read this if you want to understand how to shape our technological future and reinvigorate democracy along the way. -- Reed Hastings, co-founder and CEO of Netflix __________ A forward-thinking manifesto from three Stanford professors which reveals how big tech's obsession with optimization and efficiency has sacrificed fundamental human values and outlines steps we can take to change course, renew our democracy, and save ourselves. __________ In no more than the blink of an eye, a naive optimism about technology's liberating potential has given way to a dystopian obsession with biased algorithms, surveillance capitalism, and job-displacing robots. Yet too few of us see any alternative to accepting the onward march of technology. We have simply accepted a technological future designed for us by technologists, the venture capitalists who fund them, and the politicians who give them free rein. It doesn't need to be this way. System Error exposes the root of our current predicament: how big tech's relentless focus on optimization is driving a future that reinforces discrimination, erodes privacy, displaces workers, and pollutes the information we get. Armed with an understanding of how technologists think and exercise their power, three Stanford professors - a philosopher working at the intersection of tech and ethics, a political scientist who served under Obama, and the director of the undergraduate Computer Science program at Stanford (also an early Google engineer) - reveal how we can hold that power to account. As the dominance of big tech becomes an explosive societal conundrum, they share their provocative insights and concrete solutions to help everyone understand what is happening, what is at stake, and what we can do to control technology instead of letting it control us.
High Performance Computing: Programming and Applications presents techniques that address new performance issues in the programming of high performance computing (HPC) applications. Omitting tedious details, the book discusses hardware architecture concepts and programming techniques that are the most pertinent to application developers for achieving high performance. Even though the text concentrates on C and Fortran, the techniques described can be applied to other languages, such as C++ and Java. Drawing on their experience with chips from AMD and systems, interconnects, and software from Cray Inc., the authors explore the problems that create bottlenecks in attaining good performance. They cover techniques that pertain to each of the three levels of parallelism: Message passing between the nodes Shared memory parallelism on the nodes or the multiple instruction, multiple data (MIMD) units on the accelerator Vectorization on the inner level After discussing architectural and software challenges, the book outlines a strategy for porting and optimizing an existing application to a large massively parallel processor (MPP) system. With a look toward the future, it also introduces the use of general purpose graphics processing units (GPGPUs) for carrying out HPC computations. A companion website at www.hybridmulticoreoptimization.com contains all the examples from the book, along with updated timing results on the latest released processors.
Intel Xeon Phi Processor High Performance Programming is an all-in-one source of information for programming the Second-Generation Intel Xeon Phi product family also called Knights Landing. The authors provide detailed and timely Knights Landingspecific details, programming advice, and real-world examples. The authors distill their years of Xeon Phi programming experience coupled with insights from many expert customers - Intel Field Engineers, Application Engineers, and Technical Consulting Engineers - to create this authoritative book on the essentials of programming for Intel Xeon Phi products. Intel (R) Xeon Phi (TM) Processor High-Performance Programming is useful even before you ever program a system with an Intel Xeon Phi processor. To help ensure that your applications run at maximum efficiency, the authors emphasize key techniques for programming any modern parallel computing system whether based on Intel Xeon processors, Intel Xeon Phi processors, or other high-performance microprocessors. Applying these techniques will generally increase your program performance on any system and prepare you better for Intel Xeon Phi processors.
The book discusses the fundamentals of high-performance computing. The authors combine visualization, comprehensibility, and strictness in their material presentation, and thus influence the reader towards practical application and learning how to solve real computing problems. They address both key approaches to programming modern computing systems: multithreading-based parallelizing in shared memory systems, and applying message-passing technologies in distributed systems. The book is suitable for undergraduate and graduate students, and for researchers and practitioners engaged with high-performance computing systems. Each chapter begins with a theoretical part, where the relevant terminology is introduced along with the basic theoretical results and methods of parallel programming, and concludes with a list of test questions and problems of varying difficulty. The authors include many solutions and hints, and often sample code.
Few works are as timely and critical to the advancement of high performance computing than is this new up-to-date treatise on leading-edge directions of operating systems. It is a first-hand product of many of the leaders in this rapidly evolving field and possibly the most comprehensive. This new and important book masterfully presents the major alternative concepts driving the future of operating system design for high performance computing. In particular, it describes the major advances of monolithic operating systems such as Linux and Unix that dominate the TOP500 list. It also presents the state of the art in lightweight kernels that exhibit high efficiency and scalability at the loss of generality. Finally, this work looks forward to possibly the most promising strategy of a hybrid structure combining full service functionality with lightweight kernel operation. With this, it is likely that this new work will find its way on the shelves of almost everyone who is in any way engaged in the multi-discipline of high performance computing. (From the foreword by Thomas Sterling)
This book constitutes the refereed proceedings of the 4th Russian Supercomputing Days, RuSCDays 2018, held in Moscow, Russia, in September 2018.The 59 revised full papers and one revised short paper presented were carefully reviewed and selected from 136 submissions. The papers are organized in topical sections on parallel algorithms; supercomputer simulation; high performance architectures, tools and technologies.
This book presents the state-of-the-art in supercomputer simulation. It includes the latest findings from leading researchers using systems from the High Performance Computing Center Stuttgart (HLRS) in 2017. The reports cover all fields of computational science and engineering ranging from CFD to computational physics and from chemistry to computer science with a special emphasis on industrially relevant applications. Presenting findings of one of Europe's leading systems, this volume covers a wide variety of applications that deliver a high level of sustained performance.The book covers the main methods in high-performance computing. Its outstanding results in achieving the best performance for production codes are of particular interest for both scientists and engineers. The book comes with a wealth of color illustrations and tables of results.
'This colourful page-turner puts artificial intelligence into a human perspective . . . Metz explains this transformative technology and makes the quest thrilling.' Walter Isaacson, author of Steve Jobs ____________________________________________________ This is the inside story of a small group of mavericks, eccentrics and geniuses who turned Artificial Intelligence from a fringe enthusiasm into a transformative technology. It's the story of how that technology became big business, creating vast fortunes and sparking intense rivalries. And it's the story of breakneck advances that will shape our lives for many decades to come - both for good and for ill. ________________________________________________ 'One day soon, when computers are safely driving our roads and speaking to us in complete sentences, we'll look back at Cade Metz's elegant, sweeping Genius Makers as their birth story - the Genesis for an age of sentient machines.' Brad Stone, author of The Everything Store and The Upstarts 'A ringside seat at what may turn out to be the pivotal episode in human history . . . easy and fun to read . . . undeniably charming.' Forbes
A step-by-step guide that will enhance your skills in creating powerful systems to solve complex issues About This Book * Carlos R. Morrison from NASA will teach you to build a supercomputer with Raspberry Pi 3 * Deepen your understanding of setting up host nodes, configuring networks, and automating mountable drives * Learn various math, physics, and engineering applications to solve complex problems Who This Book Is For This book targets hobbyists and enthusiasts who want to explore building supercomputers with microcomputers. Researchers will also find this book useful. Prior programming knowledge is necessary; knowledge of supercomputers is not. What You Will Learn * Understand the concept of the Message Passing Interface (MPI) * Understand node networking. * Configure nodes so that they can communicate with each other via the network switch * Build a Raspberry Pi3 supercomputer. * Test the supercluster * Use the supercomputer to calculate MPI p codes. * Learn various practical supercomputer applications In Detail Author Carlos R. Morrison (Staff Scientist, NASA) will empower the uninitiated reader to quickly assemble and operate a Pi3 supercomputer in the shortest possible time. The lifeblood of a supercomputer, the MPI code, is introduced early, and sample MPI code provides additional practice opportunities for you to test the effectiveness of your creation. You will learn how to configure various nodes and switches so that they can effectively communicate with each other. By the end of this book, you will have successfully built a supercomputer and the various applications related to it. Style and approach A progressive guide that will start off with serial coding and MPI concepts, moving towards configuring a complete supercluster, and solving real world problems
"The more we know about smart and intelligent systems and their use, the more productive organizations can become, and the more quality of life will improve."-Gavriel Salvendy, President Academy of Science, Engineering and Medicine of Florida, University Distinguished Professor University of Central Florida" "Robots, drones, self-driving cars, and personal assistants are only some of the 'intelligent' and 'smart' systems which are populating our world and changing the way we use technology to carry out our everyday activities, bringing about both exciting opportunities for human-technology symbiosis, as well as compelling design and development challenges. Through a carefully selected choice of chapters, authored by top scientists in the field, this book, edited by Abbas Moallem, sheds light on fundamental aspects of intelligent and smart systems, investigating the role and impact of affective and psychophysiological computing, machine learning, cybersecurity, agent transparency, and human-agent teaming in the shaping of this new interaction paradigm, as well as the human factors involved in their application in critical domains such as health, education, and manufacturing in the emerging technological landscape."-Constantine Stephanidis, Professor of Computer Science, University of Crete, Distinguished member of Foundation for Research and Technology - Hellas (FORTH) In today's digital world, the words "smart" and intelligent" are now used to label devices, machinery, systems, and even environments. What is a "smart" system? Is "smart" synonymous with "intelligent"? If not, what does an "intelligent system" mean? Are all smart systems intelligent? This book tries to answer these questions by summarizing the existing research in various areas and providing new research findings. Smart and Intelligent Systems: The Human Elements in Artificial Intelligence, Robotics, and Cybersecurity presents new areas of smart and intelligent system design. It defines smart and intelligent systems, offers a human factors approach, discusses networking applications, and combines the human element with smart and intelligent systems. This book is perfect for engineering students in data sciences and artificial intelligence and practitioners at all levels in the fields of human factors and ergonomics, systems engineering, computer science, software engineering, and robotics.
Fifteen original contributions from experts in high-speed computation on multi-processor architectures, concurrent programming and parallel algorithms. Experts in high-speed computation agree that the rapidly growing demand for more powerful computers can only be met by a radical change in computer architecture, a change from a single serial processor to an aggregation of many processors working in parallel. At present, our knowledge about multi-processor architectures, concurrent programming or parallel algorithms is very limited. This book discusses all three subjects in relation to the HEP supercomputer that can handle multiple instruction streams and multiple data streams (MIMD). The HEP multiprocessor is an innovative general purpose computer, easy to use by anybody familiar with FORTRAN. Following a preface by the editor, the book's fifteen original contributions are divided into four sections: The HEP Architecture and Systems Software; The HEP Performance; Programming and Languages; and Applications of the HEP Computer. An appendix describes the use of monitors in FORTRAN, providing a tutorial on the barrier, self-scheduling DO loop, and Askfor monitors. J. S. Kowalik, who has contributed a chapter with S. P. Kumar on "Parallel Algorithms for Recurrence and Tridiagonal Linear Equations," is a manager in Boeing Computer Services' Artificial Intelligence Center in Seattle.MIMD Computation is included in the Scientific Computation Series, edited by Dennis Cannon.
Ecosystems and Technology: Idea Generation and Content Model Processing, presents important new innovations in the area of management and computing. Innovation is the generation and application of new ideas and skills to produce new products, processes, and services that improve economic and social prosperity. This includes management and design policy decisions and encompasses innovation research, analysis, and best practice in enterprises, public and private sector service organizations, government, regional societies and economies. The book, the first volume in the Innovation Management and Computing book series, looks at technology that improves efficiency and idea generation, including systems for business, medical/health, education, and more. The book provides detailed examples to provide readers with current issues, including Venture planning for innovations New technologies supporting innovations systems Competitive business modeling Context-driven innovation modeling The generation of ideas faster The measurement of relevant data Virtual interfaces Business intelligence and content processing Predictive modeling Haptic expression and emotion recognition innovations, with applications to neurocognitive medical science This book provides a wealth of information that will be useful for IT and business professionals, educators, and students in many fields.
This uniquely comprehensive book brings together the vast amount of technical, economic, and political information and the analyses of supercomputing that have hitherto been buried in the frequently inaccessible "gray literature." Seventy-nine distinguished participants in the second Frontiers of Supercomputing conference offer perceptive and often controversial views on the emerging computing environment in the United States. This title is part of UC Press's Voices Revived program, which commemorates University of California Press's mission to seek out and cultivate the brightest minds and give them voice, reach, and impact. Drawing on a backlist dating to 1893, Voices Revived makes high-quality, peer-reviewed scholarship accessible once again using print-on-demand technology. This title was originally published in 1994. |
You may like...
High Performance Computing in Science…
Wolfgang E. Nagel, Dietmar H. Kroener, …
Hardcover
R5,236
Discovery Miles 52 360
A Practical Approach to High-Performance…
Sergei Kurgalin, Sergei Borzunov
Hardcover
R1,982
Discovery Miles 19 820
High Performance Computing in Science…
Siegfried Wagner, Matthias Steinmetz, …
Hardcover
R4,157
Discovery Miles 41 570
Computational Science and High…
Egon Krause, Yurii I Shokin, …
Hardcover
R5,335
Discovery Miles 53 350
High Performance Computing on Vector…
Thomas Boenisch, Sunil Tiyyagura, …
Hardcover
R2,655
Discovery Miles 26 550
Guide to DataFlow Supercomputing - Basic…
Veljko Milutinovic, Jakob Salom, …
Hardcover
Supercomputers - A Key to U.S…
J.R. Kirkland, J.H. Poore
Hardcover
High Performance Computational Science…
Michael K. Ng, Andrei Doncescu, …
Hardcover
R2,782
Discovery Miles 27 820
|