![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer hardware & operating systems > Computer architecture & logic design > General
Telecommunications have underpinned social interaction and economic activity since the 19th century and have been increasingly reliant on optical fibers since their initial commercial deployment by BT in 1983. Today, mobile phone networks, data centers, and broadband services that facilitate our entertainment, commerce, and increasingly health provision are built on hidden optical fiber networks. However, recently it emerged that the fiber network is beginning to fill up, leading to the talk of a capacity crunch where the capacity still grows but struggles to keep up with the increasing demand. This book, featuring contributions by the suppliers of widely deployed simulation software and academic authors, illustrates the origins of the limited performance of an optical fiber from the engineering, physics, and information theoretic viewpoints. Solutions are then discussed by pioneers in each of the respective fields, with near-term solutions discussed by industrially based authors, and more speculative high-potential solutions discussed by leading academic groups.
Heterogeneous Systems Architecture - a new compute platform infrastructure presents a next-generation hardware platform, and associated software, that allows processors of different types to work efficiently and cooperatively in shared memory from a single source program. HSA also defines a virtual ISA for parallel routines or kernels, which is vendor and ISA independent thus enabling single source programs to execute across any HSA compliant heterogeneous processer from those used in smartphones to supercomputers. The book begins with an overview of the evolution of heterogeneous parallel processing, associated problems, and how they are overcome with HSA. Later chapters provide a deeper perspective on topics such as the runtime, memory model, queuing, context switching, the architected queuing language, simulators, and tool chains. Finally, three real world examples are presented, which provide an early demonstration of how HSA can deliver significantly higher performance thru C++ based applications. Contributing authors are HSA Foundation members who are experts from both academia and industry. Some of these distinguished authors are listed here in alphabetical order: Yeh-Ching Chung, Benedict R. Gaster, Juan Gomez-Luna, Derek Hower, Lee Howes, Shih-Hao HungThomas B. Jablin, David Kaeli,Phil Rogers, Ben Sander, I-Jui (Ray) Sung.
Programming Massively Parallel Processors: A Hands-on Approach shows both student and professional alike the basic concepts of parallel programming and GPU architecture. Various techniques for constructing parallel programs are explored in detail. Case studies demonstrate the development process, which begins with computational thinking and ends with effective and efficient parallel programs. Topics of performance, floating-point format, parallel patterns, and dynamic parallelism are covered in depth. For this new edition, the authors are updating their coverage of CUDA, including the concept of unified memory, and expanding content in areas such as threads, while still retaining its concise, intuitive, practical approach based on years of road-testing in the authors' own parallel computing courses.
The architects of today s large and complex systems all too often struggle with the lack of a consistent set of principles and practices that adequately address the entire breadth of systems architecture. The Method Framework for Engineering System Architectures (MFESA) enables system architects and process engineers to create methods for effectively and efficiently engineering high-quality architecture for systems, subsystems, and software components. Meets the Needs of Specific Projects The book begins by documenting the common challenges that must be addressed by system architecture engineering. It explores the major principles answering these challenges and forming the basis of MFESA. Next, the authors introduce MFESA, including its primary goals, inputs, tasks, outputs, and assumptions. Then they describe the fundamental concepts and terminology on which the systems architecture engineering is founded. This is followed by a description of each of the ten system architecture engineering tasks including associated goals and objectives, preconditions, inputs, steps, postconditions, work products, guidelines, and pitfalls. Finally, the book documents the relationship between quality and architecture, explains the quality model underlying MFESA, and provides a summary of MFESA method framework, as well as a list of points to remember and future directions planned for MFESA. Explains Specific Rationales Organized as a handy desk reference, this book harnesses more than 100 years of the authors combined professional experience to provide extensive guidelines, best practices, and tips on avoiding possible pitfalls. It presents a direct rationale of why steps are taken, how things can go wrong, and guidance for how and when to tailor the model for a system s specific context. CRC Press is pleased to announce that The Method Framework for Engineering System Architectures has been added to "Intel Corporation s Recommended Reading List." Intel s Recommended Reading program provides technical professionals a simple and handy reference list of what to read to stay abreast of new technologies. Dozens of industry technologists, corporate fellows, and engineers have helped by suggesting books and reviewing the list. This is the most comprehensive reading list available for professional computer developers.
This textbook, now in its sixth edition, continues to be straightforward and easy-to-read, presenting the principles of PLCs while not tying itself to one manufacturer or another. Extensive examples and chapter ending problems utilize several popular PLCs, highlighting understanding of fundamentals that can be used regardless of manufacturer. This book will help you to understand the main design characteristics, internal architecture, and operating principles of PLCs, as well as Identify safety issues and methods for fault diagnosis, testing, and debugging. New to This edition: A new chapter 1 with a comparison of relay-controlled systems, microprocessor-controlled systems, and the programmable logic controller, a discussion of PLC hardware and architecture, examples from various PLC manufacturers, and coverage of security, the IEC programming standard, programming devices and manufacturer's software More detail of programming using Sequential Function Charts Extended coverage of the sequencer More Information on fault finding, including testing inputs and outputs with an illustration of how it is done with the PLC manufacturer's software New case studies
A Practical Introduction to Enterprise Network and Security Management, Second Edition, provides a balanced understanding of introductory and advanced subjects in both computer networking and cybersecurity. Although much of the focus is on technical concepts, managerial issues related to enterprise network and security planning and design are explained from a practitioner’s perspective.
How can we provide guarantees of behaviours for autonomous systems such as driverless cars? This tutorial text, for professionals, researchers and graduate students, explains how autonomous systems, from intelligent robots to driverless cars, can be programmed in ways that make them amenable to formal verification. The authors review specific definitions, applications and the unique future potential of autonomous systems, along with their impact on safer decisions and ethical behaviour. Topics discussed include the use of rational cognitive agent programming from the Beliefs-Desires-Intentions paradigm to control autonomous systems and the role model-checking in verifying the properties of this decision-making component. Several case studies concerning both the verification of autonomous systems and extensions to the framework beyond the model-checking of agent decision-makers are included, along with complete tutorials for the use of the freely-available verifiable cognitive agent toolkit Gwendolen, written in Java.
After nearly six years as the field's leading reference, the second edition of this award-winning handbook reemerges with completely updated content and a brand new format. The Computer Engineering Handbook, Second Edition is now offered as a set of two carefully focused books that together encompass all aspects of the field. In addition to complete updates throughout the book to reflect the latest issues in low-power design, embedded processors, and new standards, this edition includes a new section on computer memory and storage as well as several new chapters on such topics as semiconductor memory circuits, stream and wireless processors, and nonvolatile memory technologies and applications.
This books focuses on recent break-throughs in the development of a variety of photonic devices, serving distances ranging from mm to many km, together with their electronic counter-parts, e.g. the drivers for lasers, the amplifiers following the detectors and most important, the relevant advanced VLSI circuits. It explains that as a consequence of the increasing dominance of optical interconnects for high performance workstation clusters and supercomputers their complete design has to be revised. This book thus covers for the first time the whole variety of interdependent subjects contributing to green photonics and electronics, serving communication and energy harvesting. Alternative approaches to generate electric power using organic photovoltaic solar cells, inexpensive and again energy efficient in production are summarized. In 2015, the use of the internet consumed 5-6% of the raw electricity production in developed countries. Power consumption increases rapidly and without some transformational change will use, by the middle of the next decade at the latest, the entire electricity production. This apocalyptic outlook led to a redirection of the focus of data center and HPC developers from just increasing bit rates and capacities to energy efficiency. The high speed interconnects are all based on photonic devices. These must and can be energy efficient but they operate in an electronic environment and therefore have to be considered in a wide scope that also requires low energy electronic devices, sophisticated circuit designs and clever architectures. The development of the next generation of high performance exaFLOP computers suffers from the same problem: Their energy consumption based on present device generations is essentially prohibitive.
With the popularity of hardware security research, several edited monograms have been published, which aim at summarizing the research in a particular field. Typically, each book chapter is a recompilation of one or more research papers, and the focus is on summarizing the state-of-the-art research. Different from the edited monograms, the chapters in this book are not re-compilations of research papers. The book follows a pedagogical approach. Each chapter has been planned to emphasize the fundamental principles behind the logic locking algorithms and relate concepts to each other using a systematization of knowledge approach. Furthermore, the authors of this book have contributed to this field significantly through numerous fundamental papers.
For any digital TV developer or manager, the maze of standards and
specifications related to MHP and OCAP is daunting-you have to
patch together pieces from several standards to gather all the
necessary knowledge you need to compete worldwide. The standards
themselves can be confusing, and contain many inconsistencies and
missing pieces. Interactive TV Standards provides a guide for
actually deploying these technologies for a broadcaster or product
and application developer.
This introductory user's guide to systems analysis and systems design focuses on building sustainable information systems to meet tomorrow's needs. It shows how practitioners can apply multiple participatory perspectives in development, so as to avoid future problems. As a practical guide, it is presented to be readily comprehensible and is organized to enable users to concentrate on their goals efficiently, and with minimum theoretical elaboration. The chapters follow the sequence involved in planning an information system, explaining key words, the time involved in each step, ending with a tutorial or exercises.
This introductory user's guide to systems analysis and systems design focuses on building sustainable information systems to meet tomorrow's needs. It shows how practitioners can apply multiple participatory perspectives in development, so as to avoid future problems. As a practical guide, it is presented to be readily comprehensible and is organized to enable users to concentrate on their goals efficiently, and with minimum theoretical elaboration. The chapters follow the sequence involved in planning an information system, explaining key words, the time involved in each step, ending with a tutorial or exercises.
This book focuses on domain-specific heterogeneous reconfigurable architectures, demonstrating for readers a computing platform which is flexible enough to support multiple standards, multiple modes, and multiple algorithms. The content is multi-disciplinary, covering areas of wireless communication, computing architecture, and circuit design. The platform described provides real-time processing capability with reasonable implementation cost, achieving balanced trade-offs among flexibility, performance, and hardware costs. The authors discuss efficient design methods for wireless communication processing platforms, from both an algorithm and architecture design perspective. Coverage also includes computing platforms for different wireless technologies and standards, including MIMO, OFDM, Massive MIMO, DVB, WLAN, LTE/LTE-A, and 5G.
System-on-chip (SoC) technology is revolutionizing the way computers are designed and used, driving down their cost and making them more pervasive than ever before. However, it's extremely challenging for designers to get their SoC designs right the first time. ARM System Architecture, Second Edition gives system designers an authoritative, inside perspective on SoC design -- and on ARM, the world's #1, fastest-growing SoC platform for mobile phones and information appliances. The insights in this book will be crucial to every system designer and ARM licensee seeking to build more effective SoC designs -- and get them to market more quickly. KEY TOPICS: In contrast to most ARM documentation, this book explains not only what ARM is, but why -- and how you can leverage it most effectively. Expert system designer and ARM specialist Steve Furber introduces the key design challenges associated with SoC systems, including memory hierarchy, caches, memory management, on-chip debug, and production test. Next, he presents state-of-the-art ARM-based solutions for each key problem. Furber reviews the entire ARM processor family, helping designers choose the most appropriate solutions; and covers both the ARM and Thumb programming models, providing real-world guidance for developing applications more quickly and effectively. The book includes a helpful review of the fundamentals of computer architecture, as well as valuable coverage of related topics such as digital signal processing and asynchronous design. MARKET:
Enterprise Architecture (EA) frameworks such as TOGAF and Zachman are still valid, but enterprise architects also need to adapt to the new reality of agile, DevOps, and overall disruption through digital transformation. This book will help do just that. The Change to Modern Enterprise Architecture will teach you how to use known frameworks in the new world of digital transformation. Over the course of the book, you'll learn how modern EA is helping drive strategic business decisions, create continuous and agile ("floating") architecture for scalability, and how to address quality and speed in architecture using and integrating DevSecOps frameworks in EA. This book is divided into three parts: the first explains what modern enterprise architecture is and why it's important to any business. It covers the different EA frameworks and explains what they are. In the second part, you will learn how to integrate modern development frameworks into EA, and why this knowledge will enable you to deftly respond to various business challenges. The final section of the book is all about scaling the enterprise using modern enterprise architecture. You will also see how the role of the enterprise architect is changing and how to remain in control of your architecture. Upon completing this book, you'll understand why the enterprise architect is no longer just a role overseeing the architecture strategy of a business, but has become more of a leader in driving engineering excellence. What You'll Learn Integrate DevSecOps as artifact to modern EA Use Enterprise Architecture to scale up your business Understand the changing role of the enterprise architect Define a floating architecture to enhance business agility Who This Book Is For Enterprise architects, IT architects, lead engineers, business architects, business leaders, product managers.
The latest edition of the definitive guide to the Ada language covers the full details of the core language Ada 2012 as updated by the 2016 ISO Corrigendum and introduces the key new features in Ada 2022. The book is in four parts. It begins by introducing the fundamental concepts for newcomers, before moving onto algorithmic aspects and then structural features such as OOP and multitasking. The fourth part gives details of the standard library and interaction with the external environment. Six complete executable programs illustrate the core features of the language in action. The book concludes with an appendix focussing on the new features in Ada 2022. These new features aid program proof and the efficient use of multicore architectures.
A comprehensive guide to understanding the standard and most recent advances in the design of reliable computer systems. It is organized into three sections, beginning with an in-depth review of existing reliability techniques and evaluation criteria for both hardware and software. also examined are the models for detecting faults and predicting failures, and the financial considerations which are inherent in the design, purchase, operation, and maintenance of a reliable system.
In the world of information technology, it is no longer the computer in the classical sense where the majority of IT applications is executed; computing is everywhere. More than 20 billion processors have already been fabricated and the majority of them can be assumed to still be operational. At the same time, virtually every PC worldwide is connected via the Internet. This combination of traditional and embedded computing creates an artifact of a complexity, heterogeneity, and volatility unmanageable by classical means. Each of our technical artifacts with a built-in processor can be seen as a ''Thing that Thinks," a term introduced by MIT's Thinglab. It can be expected that in the near future these billions of Things that Think will become an ''Internet of Things," a term originating from ETH Zurich. This means that we will be constantly surrounded by a virtual "organism" of Things that Think. This organism needs novel, adequate design, evolution, and management means which is also one of the core challenges addressed by the recent German priority research program on Organic Computing.
The dramatic increase in computer performance has been extraordinary, but not for all computations: it has key limits and structure. Software architects, developers, and even data scientists need to understand how exploit the fundamental structure of computer performance to harness it for future applications. Ideal for upper level undergraduates, Computer Architecture for Scientists covers four key pillars of computer performance and imparts a high-level basis for reasoning with and understanding these concepts: Small is fast - how size scaling drives performance; Implicit parallelism - how a sequential program can be executed faster with parallelism; Dynamic locality - skirting physical limits, by arranging data in a smaller space; Parallelism - increasing performance with teams of workers. These principles and models provide approachable high-level insights and quantitative modelling without distracting low-level detail. Finally, the text covers the GPU and machine-learning accelerators that have become increasingly important for mainstream applications.
This book discusses and compares several new trends that can be used to overcome Moore's law limitations, including Neuromorphic, Approximate, Parallel, In Memory, and Quantum Computing. The author shows how these paradigms are used to enhance computing capability as developers face the practical and physical limitations of scaling, while the demand for computing power keeps increasing. The discussion includes a state-of-the-art overview and the essential details of each of these paradigms.
Intelligent Vehicular Network and Communications: Fundamentals, Architectures and Solutions begins with discussions on how the transportation system has transformed into today's Intelligent Transportation System (ITS). It explores the design goals, challenges, and frameworks for modeling an ITS network, discussing vehicular network model technologies, mobility management architectures, and routing mechanisms and protocols. It looks at the Internet of Vehicles, the vehicular cloud, and vehicular network security and privacy issues. The book investigates cooperative vehicular systems, a promising solution for addressing current and future traffic safety needs, also exploring cooperative cognitive intelligence, with special attention to spectral efficiency, spectral scarcity, and high mobility. In addition, users will find a thorough examination of experimental work in such areas as Controller Area Network protocol and working function of On Board Unit, as well as working principles of roadside unit and other infrastructural nodes. Finally, the book examines big data in vehicular networks, exploring various business models, application scenarios, and real-time analytics, concluding with a look at autonomous vehicles.
Navigation Design and SEO for Content-Intensive Websites: A Guide for an Efficient Digital Communication presents the characteristics and principal guidelines for the analysis and design of efficient navigation and information access systems on content-intensive websites, such as magazines and other media publications. Furthermore, the book aims to present the tools of information processing, including information architecture (IA) and content categorization systems, so that such designs can ensure a good navigation experience based on the semantic relations between content items. The book also presents best practices in the design of information access systems with regard to their main structures, including search query forms and search result pages. Finally, the book describes the foundations of search engine optimization (SEO), emphasizing SEO oriented to publications focused on communication and the coverage of current affairs, including images and videos.
Information Security Science: Measuring the Vulnerability to Data Compromises provides the scientific background and analytic techniques to understand and measure the risk associated with information security threats. This is not a traditional IT security book since it includes methods of information compromise that are not typically addressed in textbooks or journals. In particular, it explores the physical nature of information security risk, and in so doing exposes subtle, yet revealing, connections between information security, physical security, information technology, and information theory. This book is also a practical risk management guide, as it explains the fundamental scientific principles that are directly relevant to information security, specifies a structured methodology to evaluate a host of threats and attack vectors, identifies unique metrics that point to root causes of technology risk, and enables estimates of the effectiveness of risk mitigation. This book is the definitive reference for scientists and engineers with no background in security, and is ideal for security analysts and practitioners who lack scientific training. Importantly, it provides security professionals with the tools to prioritize information security controls and thereby develop cost-effective risk management strategies.
Big Data: Principles and Paradigms captures the state-of-the-art research on the architectural aspects, technologies, and applications of Big Data. The book identifies potential future directions and technologies that facilitate insight into numerous scientific, business, and consumer applications. To help realize Big Data's full potential, the book addresses numerous challenges, offering the conceptual and technological solutions for tackling them. These challenges include life-cycle data management, large-scale storage, flexible processing infrastructure, data modeling, scalable machine learning, data analysis algorithms, sampling techniques, and privacy and ethical issues. |
You may like...
Clean Architecture - A Craftsman's Guide…
Robert Martin
Paperback
(1)
Grammatical and Syntactical Approaches…
Juhyun Lee, Michael J. Ostwald
Hardcover
R5,315
Discovery Miles 53 150
Best Practices and New Perspectives in…
Patricia Ordonez De Pablos, Robert Tennyson
Hardcover
R4,729
Discovery Miles 47 290
Novel Approaches to Information Systems…
Naveen Prakash, Deepika Prakash
Hardcover
R5,924
Discovery Miles 59 240
|