![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > General theory of computing > Systems analysis & design
Embedded systems are informally defined as a collection of programmable parts surrounded by ASICs and other standard components, that interact continuously with an environment through sensors and actuators. The programmable parts include micro-controllers and Digital Signal Processors (DSPs). Embedded systems are often used in life-critical situations, where reliability and safety are more important criteria than performance. Today, embedded systems are designed with an ad hoc approach that is heavily based on earlier experience with similar products and on manual design. Use of higher-level languages such as C helps structure the design somewhat, but with increasing complexity it is not sufficient. Formal verification and automatic synthesis of implementations are the surest ways to guarantee safety. Thus, the POLIS system which is a co-design environment for embedded systems is based on a formal model of computation. POLIS was initiated in 1988 as a research project at the University of California at Berkeley and, over the years, grew into a full design methodology with a software system supporting it. Hardware-Software Co-Design of Embedded Systems: The POLIS Approach is intended to give a complete overview of the POLIS system including its formal and algorithmic aspects. Hardware-Software Co-Design of Embedded Systems: The POLIS Approach will be of interest to embedded system designers (automotive electronics, consumer electronics and telecommunications), micro-controller designers, CAD developers and students.
Inthe?eldofformalmethodsincomputerscience,concurrencytheoryisreceivinga constantlyincreasinginterest.Thisisespeciallytrueforprocessalgebra.Althoughit had been originally conceived as a means for reasoning about the semantics of c- current programs, process algebraic formalisms like CCS, CSP, ACP, ?-calculus, and their extensions (see, e.g., [154,119,112,22,155,181,30]) were soon used also for comprehendingfunctionaland nonfunctionalaspects of the behaviorof com- nicating concurrent systems. The scienti?c impact of process calculi and behavioral equivalences at the base of process algebra is witnessed not only by a very rich literature. It is in fact worth mentioningthe standardizationprocedurethat led to the developmentof the process algebraic language LOTOS [49], as well as the implementation of several modeling and analysis tools based on process algebra, like CWB [70] and CADP [93], some of which have been used in industrial case studies. Furthermore, process calculi and behavioral equivalencesare by now adopted in university-levelcourses to teach the foundations of concurrent programming as well as the model-driven design of concurrent, distributed, and mobile systems. Nevertheless, after 30 years since its introduction, process algebra is rarely adopted in the practice of software development. On the one hand, its technica- ties often obfuscate the way in which systems are modeled. As an example, if a process term comprises numerous occurrences of the parallel composition operator, it is hard to understand the communicationscheme among the varioussubterms. On the other hand, process algebra is perceived as being dif?cult to learn and use by practitioners, as it is not close enough to the way they think of software systems.
Embedded Software for SoC covers all software related aspects of
SoC design:
Integrating the best aspects of the structured systems analysis and design method and the prototyping method, this work introduces a unique approach to computer systems development which is simple, flexible, thorough, and cost-effective.
The third edition of this authoritative and comprehensive handbook is the definitive work on the current state of the art of Biometric Presentation Attack Detection (PAD) - also known as Biometric Anti-Spoofing. Building on the success of the previous editions, this thoroughly updated third edition has been considerably revised to provide even greater coverage of PAD methods, spanning biometrics systems based on face, fingerprint, iris, voice, vein, and signature recognition. New material is also included on major PAD competitions, important databases for research, and on the impact of recent international legislation. Valuable insights are supplied by a selection of leading experts in the field, complete with results from reproducible research, supported by source code and further information available at an associated website. Topics and features: reviews the latest developments in PAD for fingerprint biometrics, covering recent technologies like Vision Transformers, and review of competition series; examines methods for PAD in iris recognition systems, the use of pupil size measurement or multiple spectra for this purpose; discusses advancements in PAD methods for face recognition-based biometrics, such as recent progress on detection of 3D facial masks and the use of multiple spectra with Deep Neural Networks; presents an analysis of PAD for automatic speaker recognition (ASV), including a study of the generalization to unseen attacks; describes the results yielded by key competitions on fingerprint liveness detection, iris liveness detection, and face anti-spoofing; provides analyses of PAD in finger-vein recognition, in signature biometrics, and in mobile biometrics; includes coverage of international standards in PAD and legal aspects of image manipulations like morphing.This text/reference is essential reading for anyone involved in biometric identity verification, be they students, researchers, practitioners, engineers, or technology consultants. Those new to the field will also benefit from a number of introductory chapters, outlining the basics for the most important biometrics. This text/reference is essential reading for anyone involved in biometric identity verification, be they students, researchers, practitioners, engineers, or technology consultants. Those new to the field will also benefit from a number of introductory chapters, outlining the basics for the most important biometrics.
Covering theoretical methods and computational techniques in biomolecular research, this book focuses on approaches for the treatment of macromolecules, including proteins, nucleic acids, and bilayer membranes. It uses concepts in free energy calculations, conformational analysis, reaction rates, and transition pathways to calculate and interpret biomolecular properties gleaned from computer-generated membrane simulations. It also demonstrates comparative protein structure modeling, outlines computer-aided drug design, discusses Bayesian statistics in molecular and structural biology, and examines the RISM-SCF/MCSCF approach to chemical processes in solution.
Network monitoring serves as the basis for a wide scope of network, engineering and management operations. Precise network monitoring involves inspecting every packet traversing in a network. However, this is not feasible with future high-speed networks, due to significant overheads of processing, storing, and transferring measured data. Network Monitoring in High Speed Networks presents accurate measurement schemes from both traffic and performance perspectives, and introduces adaptive sampling techniques for various granularities of traffic measurement. The techniques allow monitoring systems to control the accuracy of estimations, and adapt sampling probability dynamically according to traffic conditions. The issues surrounding network delays for practical performance monitoring are discussed in the second part of this book. Case studies based on real operational network traces are provided throughout this book. Network Monitoring in High Speed Networks is designed as a secondary text or reference book for advanced-level students and researchers concentrating on computer science and electrical engineering. Professionals working within the networking industry will also find this book useful.
Information System Development Improving Enterprise Communication are the collected proceedings of the 22nd International Conference on Information Systems Development: Improving Enterprise Communication ISD 2013 Conference, held in Seville, Spain. It follows in the tradition of previous conferences in the series in exploring the connections between industry, research and education. These proceedings represent ongoing reflections within the academic community on established information systems topics and emerging concepts, approaches and ideas. It is hoped that the papers herein contribute towards disseminating research and improving practice. The conference tracks highlighted at the 22nd International Conference on Information Systems Development (ISD 2013) were: ApplicationsData and OntologiesEnd UsersEnterprise EvolutionIndustrial cases in ISDIntelligent Business Process ManagementModel Driven Engineering in ISDNew TechnologiesProcess ManagementQuality"
Contains a disk of all the example problems included in the book Embedded systems are altering the landscape of electronics manufacturing worldwide, giving many consumer products sophisticated capabilities undreamt of even a few years ago. The explosive proliferation of built-in computers and the variety of design methods developed in both industry and academia necessitates the sort of pragmatic guidance offered in Embedded Systems Design with 8051 Microcontrollers. This enormously practical reference/text explains the developments in microcontroller technology and provides lucid instructions on its many and varied applications-focusing on the popular 8-bit microcontroller, the 8051, and the 83C552. Outlines a systematic methodology for design of small-scale, control-dominated embedded systems Including end-of-chapter problems that reinforce essential concepts and end-of-chapter references with URLs, Embedded Systems Design with 8051 Microcontrollers reviews basic concepts, from logic gates to Internet appliances considers 8051 and 83C552 microcontrollers as parallel running processors and embedded peripherals introduces a coherent taxonomy and symbols for microcontroller flags provides a succession of assembly language examples such as electromechanical and digital clocks examines digital interfacing at two hierarchical levels: interface to typical system components and interaction with the outside world covers applications of analog interfacing, from elementary forms to advanced designs for speech machines discusses serial interfaces suitable for distributed embedded systems demonstrates the transition from classical design approaches to the hardware-software codesign with case studies of a simplified EPROM programmer and an EPROM emulator and more Profusely illustrated with over 250 drawings and diagrams, this state-of-the-art resource is a must-read reference for electrical, electronics, computer, industrial, and
A reactive system is one that is in continual interaction with its environment and executes at a pace determined by that environment. Examples of reactive systems are network protocols, air-traffic control systems, industrial-process control systems etc. Reactive systems are ubiquitous and represent an important class of systems. Due to their complex nature, such systems are extremely difficult to specify and implement. Many reactive systems are employed in highly-critical applications, making it crucial that one considers issues such as reliability and safety while designing such systems. The design of reactive systems is considered to be problematic, and p.oses one of the greatest challenges in the field of system design and development. In this paper, we discuss specification-modeling methodologies for reactive systems. Specification modeling is an important stage in reactive system design where the designer specifies the desired properties of the reactive system in the form of a specification model. This specification model acts as the guidance and source for the implementation. To develop the specification model of complex systems in an organized manner, designers resort to specification modeling methodologies. In the context of reactive systems, we can call such methodologies reactive-system specification modeling methodologies.
The process of designing large real-time embedded signal processing systems is plagued by a lack of coherent specification and design methodology. A canonical waterfall design process is commonly used to specify, design, and implement these systems with commercial-off-the-shelf (COTS) multiprocessing (MP) hardware and software. Powerful frameworks exist for each individual phase of this canonical design process, but no single methodology exists which enables these frameworks to work together coherently, i.e., allowing the output of a framework used in one phase to be consumed by a different framework used in the next phase. This lack of coherence usually leads to design errors that are not caught until well in to the implementation phase. Since the cost of redesign increases as the design moves through these three stages, redesign is the most expensive if not performed until the implementation phase, thus making the current incoherent methodology costly. Specification and Design Methodology for Real-Time Embedded Systems shows how designs targeting COTS MP technologies can be improved by providing a coherent coupling between these frameworks, a quality known as "model continuity. This book presents a new specification and design methodology (SDM) which accomplishes the requirements specification, design exploration, and implementation of COTS MP-based signal processing systems by using powerful commercial frameworks that are intelligently integrated into a single domain-specific SDM. From the foreword: "This book is remarkably practical. It provides an excellent snapshot of the state-of-the-art and gives the reader a good understanding of both the fundamental challenges of specificationand design as well as a unified and quantified ability to assess a given methodology." Daniel Gajski, University of California
CHARM '97 is the ninth in a series of working conferences devoted to the development and use of formal techniques in digital hardware design and verification. This series is held in collaboration with IFIP WG 10.5. Previous meetings were held in Europe every other year.
This book introduces a new notion of replacement in maintenance and reliability theory. Replacement Overtime, where replacement is done at the first completion of a working cycle over a planned time, is a new research topic in maintenance theory and also serves to provide a fresh optimization technique in reliability engineering. In comparing replacement overtime with standard and random replacement techniques theoretically and numerically, 'Maintenance Overtime Policies in Reliability Theory' highlights the key improvements to be gained by adopting this new approach and shows how they can be applied to inspection policies, parallel systems and cumulative damage models. Utilizing the latest research in replacement overtime by internationally recognized experts, the reader will be introduced to new topics and methods, and learn how to apply this knowledge practically to actual reliability models. This book will serve as an essential guide to a new subject of study for graduate students and researchers and also provides a useful guide for reliability engineers and managers who have difficulties in maintenance of computer and production systems with random working cycles.
I am glad to see this new book on the e language and on verification. I am especially glad to see a description of the e Reuse Methodology (eRM). The main goal of verification is, after all, finding more bugs quicker using given resources, and verification reuse (module-to-system, old-system-to-new-system etc. ) is a key enabling component. This book offers a fresh approach in teaching the e hardware verification language within the context of coverage driven verification methodology. I hope it will help the reader und- stand the many important and interesting topics surrounding hardware verification. Yoav Hollander Founder and CTO, Verisity Inc. Preface This book provides a detailed coverage of the e hardware verification language (HVL), state of the art verification methodologies, and the use of e HVL as a facilitating verification tool in implementing a state of the art verification environment. It includes comprehensive descriptions of the new concepts introduced by the e language, e language syntax, and its as- ciated semantics. This book also describes the architectural views and requirements of verifi- tion environments (randomly generated environments, coverage driven verification environments, etc. ), verification blocks in the architectural views (i. e. generators, initiators, c- lectors, checkers, monitors, coverage definitions, etc. ) and their implementations using the e HVL. Moreover, the e Reuse Methodology (eRM), the motivation for defining such a gui- line, and step-by-step instructions for building an eRM compliant e Verification Component (eVC) are also discussed.
As electronic technology reaches the point where complex systems can be integrated on a single chip, and higher degrees of performance can be achieved at lower costs, designers must devise new ways to undertake the laborious task of coping with the numerous, and non-trivial, problems that arise during the conception of such systems. On the other hand, shorter design cycles (so that electronic products can fit into shrinking market windows) put companies, and consequently designers, under pressure in a race to obtain reliable products in the minimum period of time. New methodologies, supported by automation and abstraction, have appeared which have been crucial in making it possible for system designers to take over the traditional electronic design process and embedded systems is one of the fields that these methodologies are mainly targeting. The inherent complexity of these systems, with hardware and software components that usually execute concurrently, and the very tight cost and performance constraints, make them specially suitable to introduce higher levels of abstraction and automation, so as to allow the designer to better tackle the many problems that appear during their design. Advanced Techniques for Embedded Systems Design and Test is a comprehensive book presenting recent developments in methodologies and tools for the specification, synthesis, verification, and test of embedded systems, characterized by the use of high-level languages as a road to productivity. Each specific part of the design process, from specification through to test, is looked at with a constant emphasis on behavioral methodologies. Advanced Techniques for Embedded Systems Design and Test is essential reading for all researchers in the design and test communities as well as system designers and CAD tools developers.
Content distribution, i.e., distributing digital content from one node to another node or multiple nodes, is the most fundamental function of the Internet. Since Amazon's launch of EC2 in 2006 and Apple's release of the iPhone in 2007, Internet content distribution has shown a strong trend toward polarization. On the one hand, considerable investments have been made in creating heavyweight, integrated data centers ("heavy-cloud") all over the world, in order to achieve economies of scale and high flexibility/efficiency of content distribution. On the other hand, end-user devices ("light-end") have become increasingly lightweight, mobile and heterogeneous, creating new demands concerning traffic usage, energy consumption, bandwidth, latency, reliability, and/or the security of content distribution. Based on comprehensive real-world measurements at scale, we observe that existing content distribution techniques often perform poorly under the abovementioned new circumstances. Motivated by the trend of "heavy-cloud vs. light-end," this book is dedicated to uncovering the root causes of today's mobile networking problems and designing innovative cloud-based solutions to practically address such problems. Our work has produced not only academic papers published in prestigious conference proceedings like SIGCOMM, NSDI, MobiCom and MobiSys, but also concrete effects on industrial systems such as Xiaomi Mobile, MIUI OS, Tencent App Store, Baidu PhoneGuard, and WiFi.com. A series of practical takeaways and easy-to-follow testimonials are provided to researchers and practitioners working in mobile networking and cloud computing. In addition, we have released as much code and data used in our research as possible to benefit the community.
Memory Architecture Exploration for Programmable Embedded Systems
addresses efficient exploration of alternative memory
architectures, assisted by a "compiler-in-the-loop" that allows
effective matching of the target application to the
processor-memory architecture. This new approach for memory
architecture exploration replaces the traditional black-box view of
the memory system and allows for aggressive co-optimization of the
programmable processor together with a customized memory system.
Proven Patterns for Designing Evolvable High-Quality APIs--For Any Domain, Technology, or Platform APIs enable breakthrough innovation and digital transformation in organizations and ecosystems of all kinds. To create user-friendly, reliable and well-performing APIs, architects, designers, and developers need expert design guidance. This practical guide cuts through the complexity of API conversations and their message contents, introducing comprehensive guidelines and heuristics for designing APIs sustainably and specifying them clearly, for whatever technologies or platforms you use. In Patterns for API Design: Simplifying Integration with Loosely Coupled Message Exchanges, five expert architects and developers cover the entire API lifecycle, from launching projects and establishing goals through defining requirements, elaborating designs, planning evolution, and creating useful documentation. They crystallize the collective knowledge of many practitioners into 44 API design patterns, consistently explained with context, pros and cons, conceptual solutions, and concrete examples. To make their pattern language accessible, they present a domain model, a running case study, decision narratives with pattern selection options and criteria, and walkthroughs of real-world projects applying the patterns in two different industries. Identify and overcome API design challenges with patterns Size your endpoint types and operations adequately Design request and response messages and their representations Refine your message design for quality Plan to evolve your APIs Document and communicate your API contracts Combine patterns to solve real-world problems and make the right tradeoffs "This book provides a healthy mix of theory and practice, containing numerous nuggets of deep advice but never losing the big picture . . . grounded in real-world experience and documented with academic rigor applied and practitioner community feedback incorporated. I am confident that [it] will serve the community well, today and tomorrow." --Prof. Dr. Dr. h. c. Frank Leymann, Managing Director, Institute of Architecture of Application Systems, University of Stuttgart
The area of hybrid dynamical systems (HDS) represents a difficult and exciting challenge to control engineers and is referred to as "the control theory of tomorrow" because of its future potential for solving problems. This relatively new discipline bridges control engineering, mathematics, and computer science. There is now an emerging literature on this topic describing a number of mathematical models, heuristic algorithms, and stability criteria. However, presently there is no systematic theory of HDS. "Hybrid Dynamical Systems" focuses on a comprehensive development of HDS theory and integrates results established by the authors. The work is a self-contained informative text/reference, covering several theoretically interesting and practically significant problems concerning the use of switched controllers and examining the sensor scheduling problem. The emphasis is on classes of uncertain systems as models for HDS. Features and topics: * Focuses on the design of robust HDS in a logical and clear manner * Applies the hybrid control systems framework to two classical robust control problems: design of an optimal stable controller for a linear system and simultaneous stabilization of a collection of plants * Presents a detailed treatment of stability and H-infinity control problems for a class of HDS * Covers recent original results with complete mathematically rigorous proofs Researchers and postgraduate students in control engineering, applied mathematics, and theoretical computer science will find this book covers the latest results on this important area of research. Advanced engineering practitioners and applied researchers working in areas of controlengineering, signal processing, communications, and fault detection will find this book an up-to-date resource.
In recent years, the use of technology for the purposes of
improving and enriching traditional instructional practices has
received a great deal of attention. However, few works have
explicitly examined cognitive, psychological, and educational
principles on which technology-supported learning environments are
based. This volume attempts to cover the need for a thorough
theoretical analysis and discussion of the principles of system
design that underlie the construction of technology-enhanced
learning environments. It presents examples of technology-supported
learning environments that cover a broad range of content domains,
from the physical sciences and mathematics to the teaching of
language and literacy.
This book focuses on design in the domain of human-computer
interaction. Including a broad sampling of case studies as well as
narrower theoretical or empirical studies, it includes
consideration of educational uses of design rationale, methods for
teaching it in industry, and applications to a variety of software
and user interface/application domains. The volume promises to be
the largest collection of work on design rationale ever assembled,
and thereby to energize the considerable, widespread interest in
this topic. It will also act as a focus for the existing but
scattered work in this domain.
This book focuses on design in the domain of human-computer interaction. Including a broad sampling of case studies as well as narrower theoretical or empirical studies, it includes consideration of educational uses of design rationale, methods for teaching it in industry, and applications to a variety of software and user interface/application domains. The volume promises to be the largest collection of work on design rationale ever assembled, and thereby to energize the considerable, widespread interest in this topic. It will also act as a focus for the existing but scattered work in this domain.
This work is Volume II of a two-volume monograph on the theory of deterministic parsing of context-free grammars. Volume I, "Languages and Parsing" (Chapters 1 to 5), was an introduction to the basic concepts of formal language theory and context-free parsing. Volume II (Chapters 6 to 10) contains a thorough treat ment of the theory of the two most important deterministic parsing methods: LR(k) and LL(k) parsing. Volume II is a continuation of Volume I; together these two volumes form an integrated work, with chapters, theorems, lemmas, etc. numbered consecutively. Volume II begins with Chapter 6 in which the classical con structions pertaining to LR(k) parsing are presented. These include the canonical LR(k) parser, and its reduced variants such as the LALR(k) parser and the SLR(k) parser. The grammarclasses for which these parsers are deterministic are called LR(k) grammars, LALR(k) grammars and SLR(k) grammars; properties of these grammars are also investigated in Chapter 6. A great deal of attention is paid to the rigorous development of the theory: detailed mathematical proofs are provided for most of the results presented."
This handbook distils the wealth of expertise and knowledge from a large community of researchers and industrial practitioners in Software Product Lines (SPLs) gained through extensive and rigorous theoretical, empirical, and applied research. It is a timely compilation of well-established and cutting-edge approaches that can be leveraged by those facing the prevailing and daunting challenge of re-engineering their systems into SPLs. The selection of chapters provides readers with a wide and diverse perspective that reflects the complementary and varied expertise of the chapter authors. This perspective covers the re-engineering processes, from planning to execution. SPLs are families of systems that share common assets, allowing a disciplined software reuse. The adoption of SPL practices has shown to enable significant technical and economic benefits for the companies that employ them. However, successful SPLs rarely start from scratch, but instead, they usually start from a set of existing systems that must undergo well-defined re-engineering processes to unleash new levels of productivity and competitiveness. Practitioners will benefit from the lessons learned by the community, captured in the array of methodological and technological alternatives presented in the chapters of the handbook, and will gain the confidence for undertaking their own re-engineering challenges. Researchers and educators will find a valuable single-entry point to quickly become familiar with the state-of-the-art on the topic and the open research opportunities; including undergraduate, graduate students, and R&D engineers who want to have a comprehensive understanding of techniques in reverse engineering and re-engineering of variability-rich software systems.
Model checking is a powerful approach for the formal verification of software. When applicable, it automatically provides complete proofs of correctness, or explains, via counter-examples, why a system is not correct.This book provides a basic introduction to this new technique. The first part describes in simple terms the theoretical basis of model checking: transition systems as a formal model of systems, temporal logic as a formal language for behavioral properties, and model-checking algorithms. The second part explains how to write rich and structured temporal logic specifications in practice, while the third part surveys some of the major model checkers available. |
You may like...
Presumptive Design - Design Provocations…
Leo Frishberg, Charles Lambdin
Paperback
R1,229
Discovery Miles 12 290
Cases on Lean Thinking Applications in…
Eduardo Guilherme Satolo, Robisom Damasceno Calado
Hardcover
R5,991
Discovery Miles 59 910
Advances in Non-volatile Memory and…
Yoshio Nishi, Blanka Magyari-Kope
Paperback
R4,593
Discovery Miles 45 930
Implementing Data Analytics and…
Chintan Bhatt, Neeraj Kumar, …
Hardcover
R5,931
Discovery Miles 59 310
|