Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > Computer hardware & operating systems > Computer architecture & logic design > General
A cutting-edge guide to the theory and practice of high-speed digital system design An understanding of high-speed interconnect phenomena is essential for digital designers who must deal with the challenges posed by the ever-increasing operating speeds of today’s microprocessors. This book provides a much-needed, practical guide to the state of the art of modern digital system design, combining easily accessible explanations with immensely useful problem-solving strategies. Written by three leading Intel engineers, High-Speed Digital System Design clarifies difficult and often neglected topics involving the effects of high frequencies on digital buses and presents a variety of proven techniques and application examples. Extensive appendices, formulas, modeling techniques as well as hundreds of figures are also provided.
This book constitutes the refereed proceedings of the 14th International Conference on Theory and Applications of Satisfiability Testing, SAT 2011, held in Ann Arbor, MI, USA in June 2011. The 25 revised full papers presented together with abstracts of 2 invited talks and 10 poster papers were carefully reviewed and selected from 57 submissions. The papers are organized in topical sections on complexity analysis, binary decision diagrams, theoretical analysis, extraction of minimal unsatisfiable subsets, SAT algorithms, quantified Boolean formulae, model enumeration and local search, and empirical evaluation.
This book constitutes the refereed proceedings of the 7th International Workshop on OpenMP, IWOMP 2011, held in Chicago, IL, USA in June 2011. The 13 revised full papers presented were carefully reviewed and selected for inclusion in the proceedings. The papers are organized in topical sections on using OpenMP with application, tools for OpenMP, extensions of OpenMP, and implementation and performance.
This book constitutes the refereed proceedings of the Second International Conference on Advances in Power Electronics and Instrumentation Engineering, PEIE 2011, held at Nagpur, India, in April 2011. The 9 revised full papers presented together with 4 short papers and 7 poster papers were carefully reviewed and selected from numerous submissions. The papers address current issues in the field of power electronics, communication engineering, instrumentation engineering, digital electronics, electrical power engineering, electrical machines, information technology, control systems, and the like.
Transactions on HiPEAC aims at the timely dissemination of research contributions in computer architecture and compilation methods for high-performance embedded computer systems. Recognizing the convergence of embedded and general-purpose computer systems, this journal publishes original research on systems targeted at specific computing tasks as well as systems with broad application bases. The scope of the journal therefore covers all aspects of computer architecture, code generation and compiler optimization methods of interest to researchers and practitioners designing future embedded systems. This third issue contains 14 papers carefully reviewed and selected out of numerous submissions and is divided into four sections. The first section contains the top four papers from the Third International Conference on High-Performance Embedded Architectures and Compilers, HiPEAC 2008, held in Goeteborg, Sweden, in January 2008. The second section consists of four papers from the 8th MEDEA Workshop held in conjunction with PACT 2007 in Brasov, Romania, in September 2007. The third section contains two regular papers and the fourth section provides a snapshot from the First Workshop on Programmability Issues for Multicore Computers, MULTIPROG, held in conjunction with HiPEAC 2008.
John Maeda is one of the world's preeminent thinkers on technology and design, and in How to Speak Machine, he offers a set of simple laws that govern not only the computers of today, but the unimaginable machines of the future. Machines are already more powerful than we can comprehend, and getting more powerful at an exponential pace. Once set in motion, algorithms never tire. And when a program's size, speed and endlessness combine with its ability to learn and transform itself, the outcome can be unpredictable and dangerous. Take the seemingly instant transformation of Microsoft's chatbot into a hate-spewing racist, or how crime-predicting algorithms reinforce racial bias. How To Speak Machine provides a coherent framework for today's product designers, business leaders and policymakers to grasp this brave new world. Drawing on his wide-ranging experience from engineering to computer science to design, Maeda shows how businesses and individuals can identify opportunities afforded by technology to make world-changing and inclusive products while avoiding the pitfalls inherent to the medium.
This book contains the proceedings of the 6th International Workshop on OpenMP held in Tsukuba City, Japan, in June 2010. The International Wo- shop on OpenMP is an annual series of workshops dedicated to the promotion andadvancementofallaspectsfocusing onparallelprogrammingwithOpenMP. OpenMP is now a major programming model for shared memory systems from multi-core machines to large-scale servers. Recently, new ideas and ch- lengeshavebeenproposedtoextendOpenMPframeworktosupportaccelerators and also to exploit other forms of parallelism beyond loop-level parallelism. The workshop serves as a forum to present the latest research ideas and results related to this shared memory programming model. It also o?ers the opportunity to interact with OpenMP users, developers and the people working on the next release of the speci?cation. In response to the Call-for-Papers for the technical program, the Program Committee received a total of 23 submissions from all over the world including Asia, USA and Europe, and all submissions were carefully refereed in a rigorous processwhichrequiredatleastthreereviewsforeachpaper,usingtheEasyChair conferencesystem.The?naldecisionswerecollectivelymadeinMarch2010. Due to time and space limitations for the workshop and proceedings, only 13 papers could be selected for presentation and inclusion in the proceedings. We believe we havechosen a diverse, high-quality set of papers, re?ecting a stimulating and enjoyable workshop. Finally, we would like to thank all authors, referees,and committee members for their outstanding contributions which have ensured a continuation of the high quality of IWOMP workshops.
Over the past 20 years, software architectures have significantly contributed to the development of complex and distributed systems. Nowadays, it is recognized that one of the critical problems in the design and development of any complex software system is its architecture, i.e. the organization of its architectural elements. Software Architecture presents the software architecture paradigms based on objects, components, services and models, as well as the various architectural techniques and methods, the analysis of architectural qualities, models of representation of architectural templates and styles, their formalization, validation and testing and finally the engineering approach in which these consistent and autonomous elements can be tackled.
How a computational framework can account for the successes and failures of human cognition At the heart of human intelligence rests a fundamental puzzle: How are we incredibly smart and stupid at the same time? No existing machine can match the power and flexibility of human perception, language, and reasoning. Yet, we routinely commit errors that reveal the failures of our thought processes. What Makes Us Smart makes sense of this paradox by arguing that our cognitive errors are not haphazard. Rather, they are the inevitable consequences of a brain optimized for efficient inference and decision making within the constraints of time, energy, and memory-in other words, data and resource limitations. Framing human intelligence in terms of these constraints, Samuel Gershman shows how a deeper computational logic underpins the "stupid" errors of human cognition. Embarking on a journey across psychology, neuroscience, computer science, linguistics, and economics, Gershman presents unifying principles that govern human intelligence. First, inductive bias: any system that makes inferences based on limited data must constrain its hypotheses in some way before observing data. Second, approximation bias: any system that makes inferences and decisions with limited resources must make approximations. Applying these principles to a range of computational errors made by humans, Gershman demonstrates that intelligent systems designed to meet these constraints yield characteristically human errors. Examining how humans make intelligent and maladaptive decisions, What Makes Us Smart delves into the successes and failures of cognition.
Architectural design is a crucial first step in developing complex software intensive systems. Early design decisions establish the structures necessary for achieving broad systemic properties. However, today's organizations lack synergy between software their development processes and technological methodologies. Providing a thorough treatment of the latest theory and best practices, Architecting Software Intensive Systems: A Practitioner's Guide explains: How and when to design architectures How to weave architecture design processes into existing development processes What to do with architecture design artifacts once created The first section establishes key concepts in architectural design for software intensive systems, including architectural drivers, structures, and fundamental guidance for architectural design. The book goes on to describe the industry tested Architecture Centric Design Method. Each stage of the method is explained and the book provides all of the supporting templates and checklists. The last section discusses practical matters, including how to adopt disciplined architectural design practices into existing organizational development processes. With the principled understanding of design provided by this book, architects can temper their visceral instinct to react and be better prepared to address a broader range of design problems regardless of business context or their domain experience.
This book describes the first comprehensive approach to the optimization of interconnect architectures in 3D systems on chips (SoCs), specially addressing the challenges and opportunities arising from heterogeneous integration. Readers learn about the physical implications of using heterogeneous 3D technologies for SoC integration, while also learning to maximize the 3D-technology gains, through a physical-effect-aware architecture design. The book provides a deep theoretical background covering all abstraction-levels needed to research and architect tomorrow's 3D-integrated circuits, an extensive set of optimization methods (for power, performance, area, and yield), as well as an open-source optimization and simulation framework for fast exploration of novel designs.
This book is designed to help requirements engineers prepare for the Certified Professional for Requirements Engineering Foundation Level exam as defined by the IREB. Requirements engineering tasks have become increasingly complex. In order to ensure a high level of knowledge and competency among requirements engineers, the International Requirements Engineering Board (IREB) developed a standardized qualification called the Certified Professional for Requirements Engineering (CPRE). The certification defines the practical skills of a requirements engineer on various training levels. This book is designed for self-study and covers the curriculum for the Certified Professional for Requirements Engineering Foundation Level exam as defined by the IREB. The 2nd edition has been thoroughly revised and is aligned with the curriculum Version 2.2 of the IREB. In addition, some minor corrections to the 1st edition have been included. About IREB: The mission of the IREB is to contribute to the standardization of further education in the fields of business analysis and requirements engineering by providing syllabi and examinations, thereby achieving a higher level of applied requirements engineering. The IRE Board is comprised of a balanced mix of independent, internationally recognized experts in the fields of economy, consulting, research, and science. The IREB is a non-profit corporation. For more information visit www.certified-re.com.
The text covers important algorithm design techniques, such as greedy algorithms, dynamic programming, and divide-and-conquer, and gives applications to contemporary problems. Techniques including Fast Fourier transform, KMP algorithm for string matching, CYK algorithm for context free parsing and gradient descent for convex function minimization are discussed in detail. The book's emphasis is on computational models and their effect on algorithm design. It gives insights into algorithm design techniques in parallel, streaming and memory hierarchy computational models. The book also emphasizes the role of randomization in algorithm design, and gives numerous applications ranging from data-structures such as skip-lists to dimensionality reduction methods.
Creativity in Computing and DataFlow Supercomputing, the latest release in the Advances in Computers series published since 1960, presents detailed coverage of innovations in computer hardware, software, theory, design, and applications. In addition, it provides contributors with a medium in which they can explore topics in greater depth and breadth than journal articles typically allow. As a result, many articles have become standard references that continue to be of significant, lasting value in this rapidly expanding field.
Computational Frameworks: Systems, Models and Applications provides an overview of advanced perspectives that bridges the gap between frontline research and practical efforts. It is unique in showing the interdisciplinary nature of this area and the way in which it interacts with emerging technologies and techniques. As computational systems are a dominating part of daily lives and a required support for most of the engineering sciences, this book explores their usage (e.g. big data, high performance clusters, databases and information systems, integrated and embedded hardware/software components, smart devices, mobile and pervasive networks, cyber physical systems, etc.).
This book presents the different challenges of secure processor architecture design for architects working in industry who want to add security features to their designs as well as graduate students interested in research on architecture and hardware security. It educates readers about how the different challenges have been solved in the past and what are the best practices, i.e., the principles, for design of new secure processor architectures. Based on the careful review of past work by many computer architects and security researchers, readers also will come to know the five basic principles needed for secure processor architecture design. The book also presents existing research challenges and potential new research directions. Finally, it presents numerous design suggestions, as well as discussing pitfalls and fallacies that designers should avoid. With growing interest in computer security and the protection of the code and data which execute on commodity computers, the amount of hardware security features in today's processors has increased significantly over the recent years. No longer of just academic interest, security features inside processors have been embraced by industry as well, with a number of commercial secure processor architectures available today. This book gives readers insights into the principles behind the design of academic and commercial secure processor architectures. Secure processor architecture research is concerned with exploring and designing hardware features inside computer processors, features which can help protect confidentiality and integrity of the code and data executing on the processor. Unlike traditional processor architecture research that focuses on performance, efficiency, and energy as the first-order design objectives, secure processor architecture design has security as the first-order design objective (while still keeping the others as important design aspects that need to be considered).
With recent changes in multicore and general-purpose computing on graphics processing units, the way parallel computers are used and programmed has drastically changed. It is important to provide a comprehensive study on how to use such machines written by specialists of the domain. The book provides recent research results in high-performance computing on complex environments, information on how to efficiently exploit heterogeneous and hierarchical architectures and distributed systems, detailed studies on the impact of applying heterogeneous computing practices to real problems, and applications varying from remote sensing to tomography. The content spans topics such as Numerical Analysis for Heterogeneous and Multicore Systems; Optimization of Communication for High Performance Heterogeneous and Hierarchical Platforms; Efficient Exploitation of Heterogeneous Architectures, Hybrid CPU+GPU, and Distributed Systems; Energy Awareness in High-Performance Computing; and Applications of Heterogeneous High-Performance Computing. Covers cutting-edge research in HPC on complex environments, following an international collaboration of members of the ComplexHPC Explains how to efficiently exploit heterogeneous and hierarchical architectures and distributed systems Twenty-three chapters and over 100 illustrations cover domains such as numerical analysis, communication and storage, applications, GPUs and accelerators, and energy efficiency
John Maeda is one of the world's preeminent thinkers on technology and design, and in How to Speak Machine, he offers a set of simple laws that govern not only the computers of today, but the unimaginable machines of the future. Machines are already more powerful than we can comprehend, and getting more powerful at an exponential pace. Once set in motion, algorithms never tire. And when a program's size, speed and endlessness combine with its ability to learn and transform itself, the outcome can be unpredictable and dangerous. Take the seemingly instant transformation of Microsoft's chatbot into a hate-spewing racist, or how crime-predicting algorithms reinforce racial bias. How To Speak Machine provides a coherent framework for today's product designers, business leaders and policymakers to grasp this brave new world. Drawing on his wide-ranging experience from engineering to computer science to design, Maeda shows how businesses and individuals can identify opportunities afforded by technology to make world-changing and inclusive products while avoiding the pitfalls inherent to the medium.
Electrical Engineering Modeling for Reliability Analysis Markov Modeling for Reliability, Maintainability, Safety, and Supportability Analyses of Complex Computer Systems IEEE Press Series on Engineering of Complex Computer Systems Phillip A. Laplante and Alexander D. Stoyen, Series Editors Markov modeling has long been accepted as a fundamental and powerful technique for the fault tolerance analysis of mission-critical applications. However, the elaborate computations required have often made Markov modeling too time-consuming to be of practical use on these complex systems. With this hands-on tool, designers can use the Markov modeling technique to analyze the safety, reliability, maintainability, and cost-effectiveness factors in the full range of complex systems in use today. Featuring groundbreaking simulation software and a comprehensive reference manual, Modeling for Reliability Analysis helps system designers surmount the mathematical computations that have previously prevented effective reliability analysis. The text and software compose a valuable self-study tool that is complete with detailed explanations, examples, and a library of Markov models that can be used for experiments and as derivations for new simulation models. The book details how these analyses are conducted, while providing hands-on instructions on how to develop reliability models for the full range of system configurations. Computer-Aided Rate Modeling and Simulation (CARMS) software is an integrated modeling tool that includes a diagram-based environment for model setup, a spreadsheet-like interface for data entry, an expert system link for automatic model construction, and an interactive graphic interfacefor displaying simulation results.
Kickstart your emotion analysis journey with this hands-on, step-by-step guide to data science success Key Features * Discover the ins and outs of the end-to-end emotional analysis workflow * Explore the use of various ML models to derive meaningful insights from all sorts of data * Hone your craft by building and tweaking complex emotion analysis models in practical projects Book Description The AI winter has long thawed, but many organizations are still failing to harness the power of machine learning (ML). If you want to tap that potential and add value to your own business with cutting-edge emotion analysis, you've found what you need in this trusty guide. In Machine Learning for Emotion Analysis, you'll take your foundational data science skills and grow them in the exciting realm of emotion analysis. With its practical approach, you'll be equipped with everything you need to give your company a clear insight into what your customers are thinking. This no-nonsense guide jumps right into the practicalities of emotion analysis, teaching you how to preprocess data, build a serviceable dataset, and ensure top-notch data quality. Once you're set up for success, we get hands-on with complex ML techniques. This is where you go from the intermediate to the advanced, covering deep neural networks, support vector machines, conditional probabilities, and more, as you experience the full breadth of possibilities with emotion analysis. The book finally rounds out with a couple of in-depth use cases - a sort of sandbox for you to experiment with your newly acquired skill set. By the end of this book, you'll be ready to present yourself as a valuable asset to any organization that takes data science seriously. What you will learn * Distinguish between sentiment analysis and emotion analysis * Master the art of data preprocessing and ensure high-quality input * Expand your use of data sources through data transformation * Build models that employ cutting-edge deep learning techniques * Discover how best to tune your models' hyperparameters * Explore the use of KNN, SVM, and DNNs for advanced use cases * Build APIs and integrate your models into existing solutions * Practice your new skills by working on real-world scenarios Who This Book Is For This book is for data scientists and Python developers who want to gain insights into what people are saying about their product, company, brand, governorship, and more. Basic knowledge of machine learning and Python programming knowledge is necessary to grasp the concepts covered.
High Performance Parallelism Pearls Volume 2 offers another set of examples that demonstrate how to leverage parallelism. Similar to Volume 1, the techniques included here explain how to use processors and coprocessors with the same programming - illustrating the most effective ways to combine Xeon Phi coprocessors with Xeon and other multicore processors. The book includes examples of successful programming efforts, drawn from across industries and domains such as biomed, genetics, finance, manufacturing, imaging, and more. Each chapter in this edited work includes detailed explanations of the programming techniques used, while showing high performance results on both Intel Xeon Phi coprocessors and multicore processors. Learn from dozens of new examples and case studies illustrating "success stories" demonstrating not just the features of Xeon-powered systems, but also how to leverage parallelism across these heterogeneous systems.
System-on-chip (SoC) technology is revolutionizing the way computers are designed and used, driving down their cost and making them more pervasive than ever before. However, it's extremely challenging for designers to get their SoC designs right the first time. ARM System Architecture, Second Edition gives system designers an authoritative, inside perspective on SoC design -- and on ARM, the world's #1, fastest-growing SoC platform for mobile phones and information appliances. The insights in this book will be crucial to every system designer and ARM licensee seeking to build more effective SoC designs -- and get them to market more quickly. KEY TOPICS: In contrast to most ARM documentation, this book explains not only what ARM is, but why -- and how you can leverage it most effectively. Expert system designer and ARM specialist Steve Furber introduces the key design challenges associated with SoC systems, including memory hierarchy, caches, memory management, on-chip debug, and production test. Next, he presents state-of-the-art ARM-based solutions for each key problem. Furber reviews the entire ARM processor family, helping designers choose the most appropriate solutions; and covers both the ARM and Thumb programming models, providing real-world guidance for developing applications more quickly and effectively. The book includes a helpful review of the fundamentals of computer architecture, as well as valuable coverage of related topics such as digital signal processing and asynchronous design. MARKET:
This book describes the Splash 2 computing system as designed and
built at the Supercomputing Research Center. This is a novel
attached processor using Xilinx 4010 FPGAs as its processing
elements and whose application programming language is VHDL. This
is the first publication that details the complete Splash 2 project
-- the hardware and software systems, the architecture and their
implementations, and the design process by which the architecture
evolved from an earlier version machine. This text allows you to
understand why the machine has been engineered in the way it has.
In addition to the description of the machine, several applications
are described in detail, permitting the reader to gain an
understanding of the capabilities and the limitations of this kind
of computing device. |
You may like...
Discoverability in Digital Repositories…
Liz Woolcott, Ali Shiri
Paperback
R971
Discovery Miles 9 710
Network Evolution and Applications
Ciro Rodriguez, Vikas Kumar Jha, …
Hardcover
R1,928
Discovery Miles 19 280
Enterprise Level Security 1 & 2
Kevin Foltz, William R. Simpson
Paperback
R1,421
Discovery Miles 14 210
IoT and AI Technologies for Sustainable…
Abid Hussain, Garima Tyagi, …
Hardcover
R3,056
Discovery Miles 30 560
The Adventurous and Practical Journey to…
FANAP Studies & Research Center
Hardcover
R3,029
Discovery Miles 30 290
Edge-AI in Healthcare - Trends and…
Sonali Vyas, Akanksha Upadhyaya, …
Hardcover
R2,644
Discovery Miles 26 440
|