![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > General theory of computing > Systems analysis & design
This book analyzes the causes of failures in computing systems, their consequences, as weIl as the existing solutions to manage them. The domain is tackled in a progressive and educational manner with two objectives: 1. The mastering of the basics of dependability domain at system level, that is to say independently ofthe technology used (hardware or software) and of the domain of application. 2. The understanding of the fundamental techniques available to prevent, to remove, to tolerate, and to forecast faults in hardware and software technologies. The first objective leads to the presentation of the general problem, the fault models and degradation mechanisms wh ich are at the origin of the failures, and finally the methods and techniques which permit the faults to be prevented, removed or tolerated. This study concerns logical systems in general, independently of the hardware and software technologies put in place. This knowledge is indispensable for two reasons: * A large part of a product' s development is independent of the technological means (expression of requirements, specification and most of the design stage). Very often, the development team does not possess this basic knowledge; hence, the dependability requirements are considered uniquely during the technological implementation. Such an approach is expensive and inefficient. Indeed, the removal of a preliminary design fault can be very difficult (if possible) if this fault is detected during the product's final testing.
For courses in engineering and technical management System architecture is the study of early decision making in complex systems. This text teaches how to capture experience and analysis about early system decisions, and how to choose architectures that meet stakeholder needs, integrate easily, and evolve flexibly. With case studies written by leading practitioners, from hybrid cars to communications networks to aircraft, this text showcases the science and art of system architecture.
Behavioral Intervals in Embedded Software introduces a
comprehensive approach to timing, power, and communication analysis
of embedded software processes. Embedded software timing, power and
communication are typically not unique but occur in intervals which
result from data dependent behavior, environment timing and target
system properties.
Written by an industry insider with state of the art research at their fingertips, this book describes the Radio Access Network (RAN) architecture, starting with currently deployed 4G, followed by the description of 5G requirements and why re-thinking of the RAN architecture is needed to support these. Based on these considerations, it explains how 5G network architecture, which is currently being defined, is likely to evolve. The aim is not merely to cover relevant standards and technologies as a purely academic exercise (although a significant part of the book will be dedicated to these), but to augment these by practical deployment, to illustrate why the RAN architecture is changing and where it is going. With 5G deployments on the horizon, there is a desire within companies to both re-think the RAN architecture and to change the proprietary nature of the RAN. Correspondingly, there is increased interest in academia, standards bodies and commercial entities involved in the area.
Community structure is a salient structural characteristic of
many real-world networks. Communities are generally hierarchical,
overlapping, multi-scale and coexist with other types of structural
regularities of networks. This poses major challenges for
conventional methods of community detection. This book will
comprehensively introduce the latest advances in community
detection, especially the detection of overlapping and hierarchical
community structures, the detection of multi-scale communities in
heterogeneous networks, and the exploration of multiple types of
structural regularities. These advances have been successfully
applied to analyze large-scale online social networks, such as
Facebook and Twitter. This book provides readers a convenient way
to grasp the cutting edge of community detection in complex
networks.
Java is an exciting new object-oriented technology. Hardware for supporting objects and other features of Java such as multithreading, dynamic linking and loading is the focus of this book. The impact of Java's features on micro-architectural resources and issues in the design of Java-specific architectures are interesting topics that require the immediate attention of the research community. While Java has become an important part of desktop applications, it is now being used widely in high-end server markets, and will soon be widespread in low-end embedded computing. Java Microarchitectures contains a collection of papers providing a snapshot of the state of the art in hardware support for Java. The book covers the behavior of Java applications, embedded processors for Java, memory system design, and high-performance single-chip architectures designed to execute Java applications efficiently.
system is a complex object containing a significant percentage of elec A tronics that interacts with the Real World (physical environments, humans, etc. ) through sensing and actuating devices. A system is heterogeneous, i. e., is characterized by the co-existence of a large number of components of disparate type and function (for example, programmable components such as micro processors and Digital Signal Processors (DSPs), analog components such as AID and D/A converters, sensors, transmitters and receivers). Any approach to system design today must include software concerns to be viable. In fact, it is now common knowledge that more than 70% of the development cost for complex systems such as automotive electronics and communication systems are due to software development. In addition, this percentage is increasing constantly. It has been my take for years that the so-called hardware-software co-design problem is formulated at a too low level to yield significant results in shorten ing design time to the point needed for next generation electronic devices and systems. The level of abstraction has to be raised to the Architecture-Function co-design problem, where Function refers to the operations that the system is supposed to carry out and Architecture is the set of supporting components for that functionality. The supporting components as we said above are heteroge neous and contain almost always programmable components."
Spectral Techniques in VLSI CAD have become a subject of renewed interest in the design automation community due to the emergence of new and efficient methods for the computation of discrete function spectra. In the past, spectral computations for digital logic were too complex for practical implementation. The use of decision diagrams for spectral computations has greatly reduced this obstacle allowing for the development of new and useful spectral techniques for VLSI synthesis and verification. Several new algorithms for the computation of the Walsh, Reed-Muller, arithmetic and Haar spectra are described. The relation of these computational methods to traditional ones is also provided. Spectral Techniques in VLSI CAD provides a unified formalism of the representation of bit-level and word-level discrete functions in the spectral domain and as decision diagrams. An alternative and unifying interpretation of decision diagram representations is presented since it is shown that many of the different commonly used varieties of decision diagrams are merely graphical representations of various discrete function spectra. Viewing various decision diagrams as being described by specific sets of transformation functions not only illustrates the relationship between graphical and spectral representations of discrete functions, but also gives insight into how various decision diagram types are related. Spectral Techniques in VLSI CAD describes several new applications of spectral techniques in discrete function manipulation including decision diagram minimization, logic function synthesis, technology mapping and equivalence checking. The use of linear transformations in decision diagram size reduction is described and the relationship to the operation known as spectral translation is described. Several methods for synthesizing digital logic circuits based on a subset of spectral coefficients are described. An equivalence checking approach for functional verification is described based upon the use of matching pairs of Haar spectral coefficients.
Embedded systems are informally defined as a collection of programmable parts surrounded by ASICs and other standard components, that interact continuously with an environment through sensors and actuators. The programmable parts include micro-controllers and Digital Signal Processors (DSPs). Embedded systems are often used in life-critical situations, where reliability and safety are more important criteria than performance. Today, embedded systems are designed with an ad hoc approach that is heavily based on earlier experience with similar products and on manual design. Use of higher-level languages such as C helps structure the design somewhat, but with increasing complexity it is not sufficient. Formal verification and automatic synthesis of implementations are the surest ways to guarantee safety. Thus, the POLIS system which is a co-design environment for embedded systems is based on a formal model of computation. POLIS was initiated in 1988 as a research project at the University of California at Berkeley and, over the years, grew into a full design methodology with a software system supporting it. Hardware-Software Co-Design of Embedded Systems: The POLIS Approach is intended to give a complete overview of the POLIS system including its formal and algorithmic aspects. Hardware-Software Co-Design of Embedded Systems: The POLIS Approach will be of interest to embedded system designers (automotive electronics, consumer electronics and telecommunications), micro-controller designers, CAD developers and students.
Inthe?eldofformalmethodsincomputerscience,concurrencytheoryisreceivinga constantlyincreasinginterest.Thisisespeciallytrueforprocessalgebra.Althoughit had been originally conceived as a means for reasoning about the semantics of c- current programs, process algebraic formalisms like CCS, CSP, ACP, ?-calculus, and their extensions (see, e.g., [154,119,112,22,155,181,30]) were soon used also for comprehendingfunctionaland nonfunctionalaspects of the behaviorof com- nicating concurrent systems. The scienti?c impact of process calculi and behavioral equivalences at the base of process algebra is witnessed not only by a very rich literature. It is in fact worth mentioningthe standardizationprocedurethat led to the developmentof the process algebraic language LOTOS [49], as well as the implementation of several modeling and analysis tools based on process algebra, like CWB [70] and CADP [93], some of which have been used in industrial case studies. Furthermore, process calculi and behavioral equivalencesare by now adopted in university-levelcourses to teach the foundations of concurrent programming as well as the model-driven design of concurrent, distributed, and mobile systems. Nevertheless, after 30 years since its introduction, process algebra is rarely adopted in the practice of software development. On the one hand, its technica- ties often obfuscate the way in which systems are modeled. As an example, if a process term comprises numerous occurrences of the parallel composition operator, it is hard to understand the communicationscheme among the varioussubterms. On the other hand, process algebra is perceived as being dif?cult to learn and use by practitioners, as it is not close enough to the way they think of software systems.
Embedded Software for SoC covers all software related aspects of
SoC design:
Integrating the best aspects of the structured systems analysis and design method and the prototyping method, this work introduces a unique approach to computer systems development which is simple, flexible, thorough, and cost-effective.
The third edition of this authoritative and comprehensive handbook is the definitive work on the current state of the art of Biometric Presentation Attack Detection (PAD) - also known as Biometric Anti-Spoofing. Building on the success of the previous editions, this thoroughly updated third edition has been considerably revised to provide even greater coverage of PAD methods, spanning biometrics systems based on face, fingerprint, iris, voice, vein, and signature recognition. New material is also included on major PAD competitions, important databases for research, and on the impact of recent international legislation. Valuable insights are supplied by a selection of leading experts in the field, complete with results from reproducible research, supported by source code and further information available at an associated website. Topics and features: reviews the latest developments in PAD for fingerprint biometrics, covering recent technologies like Vision Transformers, and review of competition series; examines methods for PAD in iris recognition systems, the use of pupil size measurement or multiple spectra for this purpose; discusses advancements in PAD methods for face recognition-based biometrics, such as recent progress on detection of 3D facial masks and the use of multiple spectra with Deep Neural Networks; presents an analysis of PAD for automatic speaker recognition (ASV), including a study of the generalization to unseen attacks; describes the results yielded by key competitions on fingerprint liveness detection, iris liveness detection, and face anti-spoofing; provides analyses of PAD in finger-vein recognition, in signature biometrics, and in mobile biometrics; includes coverage of international standards in PAD and legal aspects of image manipulations like morphing.This text/reference is essential reading for anyone involved in biometric identity verification, be they students, researchers, practitioners, engineers, or technology consultants. Those new to the field will also benefit from a number of introductory chapters, outlining the basics for the most important biometrics. This text/reference is essential reading for anyone involved in biometric identity verification, be they students, researchers, practitioners, engineers, or technology consultants. Those new to the field will also benefit from a number of introductory chapters, outlining the basics for the most important biometrics.
Covering theoretical methods and computational techniques in biomolecular research, this book focuses on approaches for the treatment of macromolecules, including proteins, nucleic acids, and bilayer membranes. It uses concepts in free energy calculations, conformational analysis, reaction rates, and transition pathways to calculate and interpret biomolecular properties gleaned from computer-generated membrane simulations. It also demonstrates comparative protein structure modeling, outlines computer-aided drug design, discusses Bayesian statistics in molecular and structural biology, and examines the RISM-SCF/MCSCF approach to chemical processes in solution.
Information System Development Improving Enterprise Communication are the collected proceedings of the 22nd International Conference on Information Systems Development: Improving Enterprise Communication ISD 2013 Conference, held in Seville, Spain. It follows in the tradition of previous conferences in the series in exploring the connections between industry, research and education. These proceedings represent ongoing reflections within the academic community on established information systems topics and emerging concepts, approaches and ideas. It is hoped that the papers herein contribute towards disseminating research and improving practice. The conference tracks highlighted at the 22nd International Conference on Information Systems Development (ISD 2013) were: ApplicationsData and OntologiesEnd UsersEnterprise EvolutionIndustrial cases in ISDIntelligent Business Process ManagementModel Driven Engineering in ISDNew TechnologiesProcess ManagementQuality"
Real-time systems are used in a wide range of applications, including control, sensing, multimedia, etc. Scheduling is a central problem for these computing/communication systems since it is responsible for software execution in a timely manner. This book, the second of two volumes on the subject, brings together knowledge on specific topics and discusses the recent advances for some of them. It addresses foundations as well as the latest advances and findings in real-time scheduling, giving comprehensive references to important papers, but the chapters are short and not overloaded with confusing details. Coverage includes scheduling approaches for networks and for energy autonomous systems. Other sophisticated issues, such as feedback control scheduling and probabilistic scheduling, are also addressed. This book can serve as a textbook for courses on the topic in bachelor's degrees and in more advanced master's degree programs. It also provides a reference for computer scientists and engineers involved in the design or the development of Cyber-Physical Systems which require up-to-date real-time scheduling solutions.
Network monitoring serves as the basis for a wide scope of network, engineering and management operations. Precise network monitoring involves inspecting every packet traversing in a network. However, this is not feasible with future high-speed networks, due to significant overheads of processing, storing, and transferring measured data. Network Monitoring in High Speed Networks presents accurate measurement schemes from both traffic and performance perspectives, and introduces adaptive sampling techniques for various granularities of traffic measurement. The techniques allow monitoring systems to control the accuracy of estimations, and adapt sampling probability dynamically according to traffic conditions. The issues surrounding network delays for practical performance monitoring are discussed in the second part of this book. Case studies based on real operational network traces are provided throughout this book. Network Monitoring in High Speed Networks is designed as a secondary text or reference book for advanced-level students and researchers concentrating on computer science and electrical engineering. Professionals working within the networking industry will also find this book useful.
This book introduces a new notion of replacement in maintenance and reliability theory. Replacement Overtime, where replacement is done at the first completion of a working cycle over a planned time, is a new research topic in maintenance theory and also serves to provide a fresh optimization technique in reliability engineering. In comparing replacement overtime with standard and random replacement techniques theoretically and numerically, 'Maintenance Overtime Policies in Reliability Theory' highlights the key improvements to be gained by adopting this new approach and shows how they can be applied to inspection policies, parallel systems and cumulative damage models. Utilizing the latest research in replacement overtime by internationally recognized experts, the reader will be introduced to new topics and methods, and learn how to apply this knowledge practically to actual reliability models. This book will serve as an essential guide to a new subject of study for graduate students and researchers and also provides a useful guide for reliability engineers and managers who have difficulties in maintenance of computer and production systems with random working cycles.
Contains a disk of all the example problems included in the book Embedded systems are altering the landscape of electronics manufacturing worldwide, giving many consumer products sophisticated capabilities undreamt of even a few years ago. The explosive proliferation of built-in computers and the variety of design methods developed in both industry and academia necessitates the sort of pragmatic guidance offered in Embedded Systems Design with 8051 Microcontrollers. This enormously practical reference/text explains the developments in microcontroller technology and provides lucid instructions on its many and varied applications-focusing on the popular 8-bit microcontroller, the 8051, and the 83C552. Outlines a systematic methodology for design of small-scale, control-dominated embedded systems Including end-of-chapter problems that reinforce essential concepts and end-of-chapter references with URLs, Embedded Systems Design with 8051 Microcontrollers reviews basic concepts, from logic gates to Internet appliances considers 8051 and 83C552 microcontrollers as parallel running processors and embedded peripherals introduces a coherent taxonomy and symbols for microcontroller flags provides a succession of assembly language examples such as electromechanical and digital clocks examines digital interfacing at two hierarchical levels: interface to typical system components and interaction with the outside world covers applications of analog interfacing, from elementary forms to advanced designs for speech machines discusses serial interfaces suitable for distributed embedded systems demonstrates the transition from classical design approaches to the hardware-software codesign with case studies of a simplified EPROM programmer and an EPROM emulator and more Profusely illustrated with over 250 drawings and diagrams, this state-of-the-art resource is a must-read reference for electrical, electronics, computer, industrial, and
The process of designing large real-time embedded signal processing systems is plagued by a lack of coherent specification and design methodology. A canonical waterfall design process is commonly used to specify, design, and implement these systems with commercial-off-the-shelf (COTS) multiprocessing (MP) hardware and software. Powerful frameworks exist for each individual phase of this canonical design process, but no single methodology exists which enables these frameworks to work together coherently, i.e., allowing the output of a framework used in one phase to be consumed by a different framework used in the next phase. This lack of coherence usually leads to design errors that are not caught until well in to the implementation phase. Since the cost of redesign increases as the design moves through these three stages, redesign is the most expensive if not performed until the implementation phase, thus making the current incoherent methodology costly. Specification and Design Methodology for Real-Time Embedded Systems shows how designs targeting COTS MP technologies can be improved by providing a coherent coupling between these frameworks, a quality known as "model continuity. This book presents a new specification and design methodology (SDM) which accomplishes the requirements specification, design exploration, and implementation of COTS MP-based signal processing systems by using powerful commercial frameworks that are intelligently integrated into a single domain-specific SDM. From the foreword: "This book is remarkably practical. It provides an excellent snapshot of the state-of-the-art and gives the reader a good understanding of both the fundamental challenges of specificationand design as well as a unified and quantified ability to assess a given methodology." Daniel Gajski, University of California
A reactive system is one that is in continual interaction with its environment and executes at a pace determined by that environment. Examples of reactive systems are network protocols, air-traffic control systems, industrial-process control systems etc. Reactive systems are ubiquitous and represent an important class of systems. Due to their complex nature, such systems are extremely difficult to specify and implement. Many reactive systems are employed in highly-critical applications, making it crucial that one considers issues such as reliability and safety while designing such systems. The design of reactive systems is considered to be problematic, and p.oses one of the greatest challenges in the field of system design and development. In this paper, we discuss specification-modeling methodologies for reactive systems. Specification modeling is an important stage in reactive system design where the designer specifies the desired properties of the reactive system in the form of a specification model. This specification model acts as the guidance and source for the implementation. To develop the specification model of complex systems in an organized manner, designers resort to specification modeling methodologies. In the context of reactive systems, we can call such methodologies reactive-system specification modeling methodologies.
CHARM '97 is the ninth in a series of working conferences devoted to the development and use of formal techniques in digital hardware design and verification. This series is held in collaboration with IFIP WG 10.5. Previous meetings were held in Europe every other year.
I am glad to see this new book on the e language and on verification. I am especially glad to see a description of the e Reuse Methodology (eRM). The main goal of verification is, after all, finding more bugs quicker using given resources, and verification reuse (module-to-system, old-system-to-new-system etc. ) is a key enabling component. This book offers a fresh approach in teaching the e hardware verification language within the context of coverage driven verification methodology. I hope it will help the reader und- stand the many important and interesting topics surrounding hardware verification. Yoav Hollander Founder and CTO, Verisity Inc. Preface This book provides a detailed coverage of the e hardware verification language (HVL), state of the art verification methodologies, and the use of e HVL as a facilitating verification tool in implementing a state of the art verification environment. It includes comprehensive descriptions of the new concepts introduced by the e language, e language syntax, and its as- ciated semantics. This book also describes the architectural views and requirements of verifi- tion environments (randomly generated environments, coverage driven verification environments, etc. ), verification blocks in the architectural views (i. e. generators, initiators, c- lectors, checkers, monitors, coverage definitions, etc. ) and their implementations using the e HVL. Moreover, the e Reuse Methodology (eRM), the motivation for defining such a gui- line, and step-by-step instructions for building an eRM compliant e Verification Component (eVC) are also discussed.
As electronic technology reaches the point where complex systems can be integrated on a single chip, and higher degrees of performance can be achieved at lower costs, designers must devise new ways to undertake the laborious task of coping with the numerous, and non-trivial, problems that arise during the conception of such systems. On the other hand, shorter design cycles (so that electronic products can fit into shrinking market windows) put companies, and consequently designers, under pressure in a race to obtain reliable products in the minimum period of time. New methodologies, supported by automation and abstraction, have appeared which have been crucial in making it possible for system designers to take over the traditional electronic design process and embedded systems is one of the fields that these methodologies are mainly targeting. The inherent complexity of these systems, with hardware and software components that usually execute concurrently, and the very tight cost and performance constraints, make them specially suitable to introduce higher levels of abstraction and automation, so as to allow the designer to better tackle the many problems that appear during their design. Advanced Techniques for Embedded Systems Design and Test is a comprehensive book presenting recent developments in methodologies and tools for the specification, synthesis, verification, and test of embedded systems, characterized by the use of high-level languages as a road to productivity. Each specific part of the design process, from specification through to test, is looked at with a constant emphasis on behavioral methodologies. Advanced Techniques for Embedded Systems Design and Test is essential reading for all researchers in the design and test communities as well as system designers and CAD tools developers. |
You may like...
Trends in Deep Learning Methodologies…
Vincenzo Piuri, Sandeep Raj, …
Paperback
R3,018
Discovery Miles 30 180
Modern Concurrency on Apple Platforms…
Andres Ibanez Kautsch
Paperback
Blockchain Technology for Global Social…
Jane Thomason, Sonja Bernhardt, …
Hardcover
R4,865
Discovery Miles 48 650
Big Data Governance and Perspectives in…
Sheryl Kruger Strydom, Moses Strydom
Hardcover
R4,886
Discovery Miles 48 860
Conceptual Modeling Perspectives
Jordi Cabot, Cristina Gomez, …
Hardcover
The Last Crusades - the Final Attempts…
Edwin Pears, T. A. Archer, …
Hardcover
R764
Discovery Miles 7 640
|