Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > Computer hardware & operating systems
This book-presents new methods and tools for the integration and simulation of smart devices. The design approach described in this book explicitly accounts for integration of Smart Systems components and subsystems as a specific constraint. It includes methodologies and EDA tools to enable multi-disciplinary and multi-scale modeling and design, simulation of multi-domain systems, subsystems and components at all levels of abstraction, system integration and exploration for optimization of functional and non-functional metrics. By covering theoretical and practical aspects of smart device design, this book targets people who are working and studying on hardware/software modelling, component integration and simulation under different positions (system integrators, designers, developers, researchers, teachers, students etc.). In particular, it is a good introduction to people who have interest in managing heterogeneous components in an efficient and effective way on different domains and different abstraction levels. People active in smart device development can understand both the current status of practice and future research directions. * Provides a comprehensive overview of smart systems design, focusing on design challenges and cutting-edge solutions; * Enables development of a co-simulation and co-design environment that accounts for the peculiarities of the basic subsystems and components to be integrated; * Describes development of modeling and design techniques, methods and tools that enable multi-domain simulation and optimization at various levels of abstraction and across different technological domains.
Unlike so many books that focus on how to use Linux, Linux and the
Unix Philosophy explores the "way of thinking that is Linux" and
why Linux is a superior implementation of this highly capable
operating system.
This volume presents a selection of reports from scientific projects requiring high end computing resources on the Hitachi SR8000-F1 supercomputer operated by Leibniz Computing Center in Munich. All reports were presented at the joint HLRB and KONWHIR workshop at the Technical University of Munich in October 2002. The following areas of scientific research are covered: Applied Mathematics, Biosciences, Chemistry, Computational Fluid Dynamics, Cosmology, Geosciences, High-Energy Physics, Informatics, Nuclear Physics, Solid-State Physics. Moreover, projects from interdisciplinary research within the KONWIHR framework (Competence Network for Scientific High Performance Computing in Bavaria) are also included. Each report summarizes its scientific background and discusses the results with special consideration of the quantity and quality of Hitachi SR8000 resources needed to complete the research.
This book provides the details of developing a digital pedestrian map construction system over the intermittently connected mobile network. Over the past couple of decades, countries across the world, both developing and developed, have witnessed a significant number of disasters. Thus, it has become mandatory for each of the disaster-prone countries to equip themselves with appropriate measures to cope with the challenges of providing post-disaster services. Some of the serious challenges are incapacitated communication infrastructure, unstable power supply and inaccessible road networks. Out of these challenges, the destruction of road networks, especially in developing countries, acts as a major hindrance to effective disaster management. To be more specific, the success of a disaster response operation generally depends on the speed of evacuation and transportation of adequate amount of relief materials at the right time to the disaster-affected areas. Hence, map-based navigation support is a primary requirement for post-disaster relief operations. This book also provides the solution of the two other important post-disaster management services such as situational awareness and resource allocation. Both of these services are invariably dependent on the existence of navigation support. Finally, in order to offer such services, the other challenge is to address the problem of incapacitated communication infrastructure. This book also deals with such challenges in post-disaster scenarios and develops automated post-disaster management services.
This volume gives an overview of the state-of-the-art with respect to the development of all types of parallel computers and their application to a wide range of problem areas.
This book describes simple to complex ASIC design practical scenarios using Verilog. It builds a story from the basic fundamentals of ASIC designs to advanced RTL design concepts using Verilog. Looking at current trends of miniaturization, the contents provide practical information on the issues in ASIC design and synthesis using Synopsys DC and their solution. The book explains how to write efficient RTL using Verilog and how to improve design performance. It also covers architecture design strategies, multiple clock domain designs, low-power design techniques, DFT, pre-layout STA and the overall ASIC design flow with case studies. The contents of this book will be useful to practicing hardware engineers, students, and hobbyists looking to learn about ASIC design and synthesis.
Dealing with system problems from user login failures to server
crashes--is a critical part of a system administrator's job. A down
system can cost a business thousands of dollars per minute. But
there is little or no information available on how to troubleshoot
and correct system problems; in most cases, these skills are
learned in an ad-hoc manner, usually in the pressure-cooker
environment of a crisis. This is the first book to address this
lack of information.
This book describes a new design methodology that allows optimization-based synthesis of RF systems in a hierarchical multilevel approach, in which the system is designed in a bottom-up fashion, from the device level up to the (sub)system level. At each level of the design hierarchy, the authors discuss methods that increase the design robustness and increase the accuracy and efficiency of the simulations. The methodology described enables circuit sizing and layout in a complete and automated integrated manner, achieving optimized designs in significantly less time than with traditional approaches.
This two-part book puts the spotlight on how a real-time kernel works using Micrium's C/OS-III kernel as a reference. Part I includes an overview of the operation of real-time kernels, and walks through various aspects of C/OS-III implementation and usage. Part II provides application examples (using the versatile Renesas YRDKRX62N Evaluation Board, available separately) that enable readers to rapidly develop their own prototypes. This book is written for serious embedded systems programmers, consultants, hobbyists, and students interested in understanding the inner workings of a real-time kernel. C/OS-III is not just a great learning platform, but also a full commercial-grade software package, ready to be part of a wide range of products. C/OS-III is a highly portable, ROMable, scalable, preemptive real-time, multitasking kernel designed specifically to address the demanding requirements of today 's embedded systems. C/OS-III is the successor to the highly popular C/OS-II real-time kernel but can use most of C/OS-II 's ports with minor modifications. Some of the features of C/OS-III are: Preemptive multitasking with round-robin scheduling of tasks at
the same priority
The state of the art in supercomputing is summarized in this volume. The book presents selected results of the projects of the High Performance Computing Center Stuttgart (HLRS) for the year 2001. Together these contributions provide an overview of recent developments in high performance computing and simulation. Reflecting the close cooperation of the HLRS with industry, special emphasis has been put on the industrial relevance of the presented results and methods. The book therefore becomes a collection of showcases for an innovative usage of state-of-the-art modeling, novel numerical algorithms and the use of leading edge high performance computing systems in a GRID-like environment.
This book offers readers comprehensive coverage of security policy specification using new policy languages, implementation of security policies in Systems-on-Chip (SoC) designs - current industrial practice, as well as emerging approaches to architecting SoC security policies and security policy verification. The authors focus on a promising security architecture for implementing security policies, which satisfies the goals of flexibility, verification, and upgradability from the ground up, including a plug-and-play hardware block in which all policy implementations are enclosed. Using this architecture, they discuss the ramifications of designing SoC security policies, including effects on non-functional properties (power/performance), debug, validation, and upgrade. The authors also describe a systematic approach for "hardware patching", i.e., upgrading hardware implementations of security requirements safely, reliably, and securely in the field, meeting a critical need for diverse Internet of Things (IoT) devices. Provides comprehensive coverage of SoC security requirements, security policies, languages, and security architecture for current and emerging computing devices; Explodes myths and ambiguities in SoC security policy implementations, and provide a rigorous treatment of the subject; Demonstrates a rigorous, step-by-step approach to developing a diversity of SoC security policies; Introduces a rigorous, disciplined approach to "hardware patching", i.e., secure technique for updating hardware functionality of computing devices in-field; Includes discussion of current and emerging approaches for security policy verification.
This book introduces a new level of abstraction that closes the gap between the textual specification of embedded systems and the executable model at the Electronic System Level (ESL). Readers will be enabled to operate at this new, Formal Specification Level (FSL), using models which not only allow significant verification tasks in this early stage of the design flow, but also can be extracted semi-automatically from the textual specification in an interactive manner. The authors explain how to use these verification tasks to check conceptual properties, e.g. whether requirements are in conflict, as well as dynamic behavior, in terms of execution traces.
This thesis focuses on the problem of optimizing the quality of network multimedia services. This problem spans multiple domains, from subjective perception of multimedia quality to computer networks management. The work done in this thesis approaches the problem at different levels, developing methods for modeling the subjective perception of quality based on objectively measurable parameters of the multimedia coding process as well as the transport over computer networks. The modeling of subjective perception is motivated by work done in psychophysics, while using Machine Learning techniques to map network conditions to the human perception of video services. Furthermore, the work develops models for efficient control of multimedia systems operating in dynamic networked environments with the goal of delivering optimized Quality of Experience. Overall this thesis delivers a set of methods for monitoring and optimizing the quality of multimedia services that adapt to the dynamic environment of computer networks in which they operate.
This book presents techniques necessary to predict cardiac arrhythmias, long before they occur, based on minimal ECG data. The authors describe the key information needed for automated ECG signal processing, including ECG signal pre-processing, feature extraction and classification. The adaptive and novel ECG processing techniques introduced in this book are highly effective and suitable for real-time implementation on ASICs.
This book puts the spotlight on how a real-time kernel works using Micrium s C/OS-III as a reference. The book consists of two complete parts. The first describes real-time kernels in generic terms. Part II provide examples for the reader, using Texas Instruments EVM-EVALBOT, a small, robotic evaluation board. The board is based on the Stellaris LM3S9B92 which combines the popular ARM Cortex-M3(r) architecture with Ethernet MAC+PHY, USB OTG (On-The-Go), and I2S. Together with the IAR Systems Embedded Workbench for ARM development tools, the evaluation board provides everything necessary to enable the reader to be up and running quickly, as well as a fun and educational experience, resulting in a high-level of proficiency in a short time. This book is written for serious embedded systems programmers, consultants, hobbyists, and students interested in understanding the inner workings of a real-time kernel. C/OS-III is not just a great learning platform, but also a full commercial-grade software package, ready to be part of a wide range of products. C/OS-III is a highly portable, ROMable, scalable, preemptive real-time, multitasking kernel designed specifically to address the demanding requirements of today s embedded systems. C/OS-III is the successor to the highly popular C/OS-II real-time kernel but can use most of C/OS-II s ports with minor modifications. Some of the features of C/OS-III are: Preemptive multitasking with round-robin scheduling of tasks at
the same priority
The question whether molecular primitives can prove to be real alternatives to contemporary semiconductor means or effective supplements extending greatly possibilities of information technologies is addressed. Molecular primitives and circuitry for information processing devices are also discussed. Investigations in molecular based computing devices were initiated in the early 1970s in the hopes for an increase in the integration level and processing speed. Real progress proved unfeasible into the 1980 s. However, recently, important and promising results were achieved. The elaboration of operational 160-kilobit molecular electronic memory patterned 1011 bits per square centimeter in the end of 90?'s were the first timid steps of information processing further development. Subsequent advances beyond these developments are presented and discussed. This work provides useful knowledge to anyone working in molecular based information processing.
Covering system architecture, implementation and testing, this work is written by authors who are widely experienced with cellular radio in general and with GSM in particular. It provides a structured overview to help make sense of the GSM specifications and surveys competing cellular systems such as NADC and CDMA. Practical testing applications are explored in depth and compared with similar techniques used with analogue cellular systems.
This book explores near-threshold computing (NTC), a design-space using techniques to run digital chips (processors) near the lowest possible voltage. Readers will be enabled with specific techniques to design chips that are extremely robust; tolerating variability and resilient against errors. Variability-aware voltage and frequency allocation schemes will be presented that will provide performance guarantees, when moving toward near-threshold manycore chips. * Provides an introduction to near-threshold computing, enabling reader with a variety of tools to face the challenges of the power/utilization wall; * Demonstrates how to design efficient voltage regulation, so that each region of the chip can operate at the most efficient voltage and frequency point; * Investigates how performance guarantees can be ensured when moving towards NTC manycores through variability-aware voltage and frequency allocation schemes.
This book provides a comprehensive overview of both theoretical and pragmatic aspects of resource-allocation and scheduling in multiprocessor and multicore hard-real-time systems. The authors derive new, abstract models of real-time tasks that capture accurately the salient features of real application systems that are to be implemented on multiprocessor platforms, and identify rules for mapping application systems onto the most appropriate models. New run-time multiprocessor scheduling algorithms are presented, which are demonstrably better than those currently used, both in terms of run-time efficiency and tractability of off-line analysis. Readers will benefit from a new design and analysis framework for multiprocessor real-time systems, which will translate into a significantly enhanced ability to provide formally verified, safety-critical real-time systems at a significantly lower cost.
This thesis takes an empirical approach to understanding of the behavior and interactions between the two main components of reinforcement learning: the learning algorithm and the functional representation of learned knowledge. The author approaches these entities using design of experiments not commonly employed to study machine learning methods. The results outlined in this work provide insight as to what enables and what has an effect on successful reinforcement learning implementations so that this learning method can be applied to more challenging problems.
This book highlights how real-time kernels work, using Micrium s C/OS-III as a reference. The book consists of two parts: Part I describes real-time kernels in generic terms, while Part II provides practical examples using NXP s LPC1768 Microcontroller, based on the ARM Cortex M3(rev 2) architecture. A companion evaluation board (Keil MCB1700) and IDE (Keil MDK Evaluation Version) enable the reader to quickly and easily evaluate the microcontroller, tools and RTOS. A range of examples are included, providing a unique hands-on experience, and leading to a faster and better understanding of the concepts presented in the book. This book is written for serious embedded systems programmers, consultants, hobbyists, and students interested in understanding the inner workings of a real-time kernel. C/OS-III is not just a great learning platform, but also a full commercial-grade software package, ready to be part of a wide range of products. C/OS-III is a highly portable, ROMable, scalable, preemptive real-time, multitasking kernel designed specifically to address the demanding requirements of today s embedded systems. C/OS-III is the successor to the highly popular C/OS-II real-time kernel but can use most of C/OS-II s ports with minor modifications. Some of the features of C/OS-III are: Preemptive multitasking with round-robin scheduling of tasks at
the same priority
This book describes the state-of-the art of industrial and academic research in the architectural design of heterogeneous, multi/many-core processors. The authors describe methods and tools to enable next-generation embedded and high-performance heterogeneous processors to confront cost-effectively the inevitable variations by providing Dependable-Performance: correct functionality and timing guarantees throughout the expected lifetime of a platform under thermal, power, and energy constraints. Various aspects of the reliability problem are discussed, at both the circuit and architecture level, the intelligent selection of knobs and monitors in multicore platforms, and systematic design methodologies. The authors demonstrate how new techniques have been applied in real case studies from different applications domain and report on results and conclusions of those experiments. Enables readers to develop performance-dependable heterogeneous multi/many-core architectures Describes system software designs that support high performance dependability requirements Discusses and analyzes low level methodologies to tradeoff conflicting metrics, i.e. power, performance, reliability and thermal management Includes new application design guidelines to improve performance dependability
This book provides readers with insight into an alternative approach for enhancing the reliability, security, and low power features of integrated circuit designs, related to transient faults, hardware Trojans, and power consumption. The authors explain how the addition of integrated sensors enables the detection of ionizing particles and how this information can be processed at a high layer. The discussion also includes a variety of applications, such as the detection of hardware Trojans and fault attacks, and how sensors can operate to provide different body bias levels and reduce power costs. Readers can benefit from these sensors-based approaches through designs with fast response time, non-intrusive integration on gate-level and reasonable design costs. |
You may like...
|