![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > General theory of computing > Systems analysis & design
Providing an introduction to the design of embedded microprocessor systems, this edition covers everything from the initial concept through to debugging the final result. It also includes material on DMA, interrupts and an emphasis throughout on the real-time nature of embedded systems. The book is not limited to describing any specific processor family, but covers the operation of, and interfaces to, several types of processors with an emphasis on cost and design trade-offs Included throughout the book are numerous examples, tips, and pitfalls to help readers find out how to implement faster and better design processes and avoid time-consuming and expensive mistakes. The author describes the entire process of designing circuits, and the software that controls them, assessing the system requirements, as well as testing and debugging systems. In this third edition, there is an expanded section on debug which includes avoiding common hardware, software and interrupt problems. Other added features include an expanded section on system integration and debug to address the capabilities of more recent emulators and debuggers, a section about combination microcontroller/PLD devices, and
To optimally design and manage a directory service, IS architects
and managers must understand current state-of-the-art products.
Directory Services covers Novell's NDS eDirectory, Microsoft's
Active Directory, UNIX directories and products by NEXOR, MaxWare,
Siemens, Critical Path and others. Directory design fundamentals
and products are woven into case studies of large enterprise
deployments. Cox thoroughly explores replication, security,
migration and legacy system integration and interoperability.
Business issues such as how to cost justify, plan, budget and
manage a directory project are also included. The book culminates
in a visionary discussion of future trends and emerging directory
technologies including the strategic direction of the top directory
products, the impact of wireless technology on directory enabled
applications and using directories to customize content delivery
from the Enterprise Portal.
This book describes a variety of test generation algorithms for testing crosstalk delay faults in VLSI circuits. It introduces readers to the various crosstalk effects and describes both deterministic and simulation-based methods for testing crosstalk delay faults. The book begins with a focus on currently available crosstalk delay models, test generation algorithms for delay faults and crosstalk delay faults, before moving on to deterministic algorithms and simulation-based algorithms used to test crosstalk delay faults. Given its depth of coverage, the book will be of interest to design engineers and researchers in the field of VLSI Testing.
With complex systems and complex requirements being a challenge that designers must face to reach quality results, multi-formalism modeling offers tools and methods that allow modelers to exploit the benefits of different techniques in a general framework intended to address these challenges. Theory and Application of Multi-Formalism Modeling boldly explores the importance of this topic by gathering experiences, theories, applications, and solutions from diverse perspectives of those involved with multi-formalism modeling. Professionals, researchers, academics, and students in this field will be able to critically evaluate the latest developments and future directions of multi-formalism research.
Due to the role of software systems in safety-critical applications and in the satisfaction of customers and organizations, the development of efficient software engineering is essential. Designing, Engineering, and Analyzing Reliable and Efficient Software discusses and analyzes various designs, systems, and advancements in software engineering. With its coverage on the integration of mathematics, computer science, and practices in engineering, this book highlights the importance of ensuring and maintaining reliable software and is an essential resource for practitioners, professors and students in these fields of study.
The book focuses on system dependability modeling and calculation, considering the impact of s-dependency and uncertainty. The best suited approaches for practical system dependability modeling and calculation, (1) the minimal cut approach, (2) the Markov process approach, and (3) the Markov minimal cut approach as a combination of (1) and (2) are described in detail and applied to several examples. The stringently used Boolean logic during the whole development process of the approaches is the key for the combination of the approaches on a common basis. For large and complex systems, efficient approximation approaches, e.g. the probable Markov path approach, have been developed, which can take into account s-dependencies be-tween components of complex system structures. A comprehensive analysis of aleatory uncertainty (due to randomness) and epistemic uncertainty (due to lack of knowledge), and their combination, developed on the basis of basic reliability indices and evaluated with the Monte Carlo simulation method, has been carried out. The uncertainty impact on system dependability is investigated and discussed using several examples with different levels of difficulty. The applications cover a wide variety of large and complex (real-world) systems. Actual state-of-the-art definitions of terms of the IEC 60050-192:2015 standard, as well as the dependability indices, are used uniformly in all six chapters of the book.
The objective of Developing Quality Complex Database Systems is to provide opportunities for improving today's database systems using innovative development practices, tools and techniques. Each chapter of this book will provide insight into the effective use of database technology through models, case studies or experience reports. An emphasis is placed on organizational and management issues associated with the use of such technology inclusive of lessons learned and best practices.
This textbook for a one-semester course in Digital Systems Design describes the basic methods used to develop "traditional" Digital Systems, based on the use of logic gates and flip flops, as well as more advanced techniques that enable the design of very large circuits, based on Hardware Description Languages and Synthesis tools. It was originally designed to accompany a MOOC (Massive Open Online Course) created at the Autonomous University of Barcelona (UAB), currently available on the Coursera platform. Readers will learn what a digital system is and how it can be developed, preparing them for steps toward other technical disciplines, such as Computer Architecture, Robotics, Bionics, Avionics and others. In particular, students will learn to design digital systems of medium complexity, describe digital systems using high level hardware description languages, and understand the operation of computers at their most basic level. All concepts introduced are reinforced by plentiful illustrations, examples, exercises, and applications. For example, as an applied example of the design techniques presented, the authors demonstrate the synthesis of a simple processor, leaving the student in a position to enter the world of Computer Architecture and Embedded Systems.
This book precisely formulates and simplifies the presentation of Instruction Level Parallelism (ILP) compilation techniques. It uniquely offers consistent and uniform descriptions of the code transformations involved. Due to the ubiquitous nature of ILP in virtually every processor built today, from general purpose CPUs to application-specific and embedded processors, this book is useful to the student, the practitioner and also the researcher of advanced compilation techniques. With an emphasis on fine-grain instruction level parallelism, this book will also prove interesting to researchers and students of parallelism at large, in as much as the techniques described yield insights that go beyond superscalar and VLIW (Very Long Instruction Word) machines compilation and are more widely applicable to optimizing compilers in general. ILP techniques have found wide and crucial application in Design Automation, where they have been used extensively in the optimization of performance as well as area and power minimization of computer designs.
Computational Frameworks: Systems, Models and Applications provides an overview of advanced perspectives that bridges the gap between frontline research and practical efforts. It is unique in showing the interdisciplinary nature of this area and the way in which it interacts with emerging technologies and techniques. As computational systems are a dominating part of daily lives and a required support for most of the engineering sciences, this book explores their usage (e.g. big data, high performance clusters, databases and information systems, integrated and embedded hardware/software components, smart devices, mobile and pervasive networks, cyber physical systems, etc.).
This book constitutes the refereed proceedings of the 14th IFIP WG 2.13 International Conference on Open Source Systems, OSS 2018, held in Athens, Greece, in June 2018. The 14 revised full papers and 2 short papers presented were carefully reviewed and selected from 38 submissions. The papers cover a wide range of topics in the field of free/libre open source software (FLOSS) and are organized in the following thematic sections: organizational aspects of OSS projects, OSS projects validity, mining OSS data, OSS in public administration, OSS governance, and OSS reusability.
This monograph combines the commutant lifting theorem for operator theory and the state space method from system theory to provide a unified approach for solving both stationary and nonstationary interpolation problems with norm constraints. Included are the operator-valued versions of the tangential Nevanlinna-Pick problem, the Hermite-FejA(c)r problem, the Nehari problem, the Sarason problem, and the two-sided Nudelman problem, and their nonstationary analogues. The main results concern the existence of solutions, the explicit construction of the central solutions in state space form, the maximum entropy property of the central solutions, and state space parametrizations of all solutions. Direct connections between the various interpolation problems are displayed. Applications to H infinity] control problems are presented. This monograph should appeal to a wide group of mathematicians and engineers. The material is self-contained and may be used for advanced graduate courses and seminars.
This unique volume introduces and discusses the methods of validating computer simulations in scientific research. The core concepts, strategies, and techniques of validation are explained by an international team of pre-eminent authorities, drawing on expertise from various fields ranging from engineering and the physical sciences to the social sciences and history. The work also offers new and original philosophical perspectives on the validation of simulations. Topics and features: introduces the fundamental concepts and principles related to the validation of computer simulations, and examines philosophical frameworks for thinking about validation; provides an overview of the various strategies and techniques available for validating simulations, as well as the preparatory steps that have to be taken prior to validation; describes commonly used reference points and mathematical frameworks applicable to simulation validation; reviews the legal prescriptions, and the administrative and procedural activities related to simulation validation; presents examples of best practice that demonstrate how methods of validation are applied in various disciplines and with different types of simulation models; covers important practical challenges faced by simulation scientists when applying validation methods and techniques; offers a selection of general philosophical reflections that explore the significance of validation from a broader perspective. This truly interdisciplinary handbook will appeal to a broad audience, from professional scientists spanning all natural and social sciences, to young scholars new to research with computer simulations. Philosophers of science, and methodologists seeking to increase their understanding of simulation validation, will also find much to benefit from in the text.
This book presents a design methodology that is practically applicable to the architectural design of a broad range of systems. It is based on fundamental design concepts to conceive and specify the required functional properties of a system, while abstracting from the specific implementation functions and technologies that can be chosen to build the system. Abstraction and precision are indispensable when it comes to understanding complex systems and precisely creating and representing them at a high functional level. Once understood, these concepts appear natural, self-evident and extremely powerful, since they can directly, precisely and concisely reflect what is considered essential for the functional behavior of a system. The first two chapters present the global views on how to design systems and how to interpret terms and meta-concepts. This informal introduction provides the general context for the remainder of the book. On a more formal level, Chapters 3 through 6 present the main basic design concepts, illustrating them with examples. Language notations are introduced along with the basic design concepts. Lastly, Chapters 7 to 12 discuss the more intricate basic design concepts of interactive systems by focusing on their common functional goal. These chapters are recommended to readers who have a particular interest in the design of protocols and interfaces for various systems. The didactic approach makes it suitable for graduate students who want to develop insights into and skills in developing complex systems, as well as practitioners in industry and large organizations who are responsible for the design and development of large and complex systems. It includes numerous tangible examples from various fields, and several appealing exercises with their solutions.
This book presents the latest key research into the performance and reliability aspects of dependable fault-tolerant systems and features commentary on the fields studied by Prof. Kishor S. Trivedi during his distinguished career. Analyzing system evaluation as a fundamental tenet in the design of modern systems, this book uses performance and dependability as common measures and covers novel ideas, methods, algorithms, techniques, and tools for the in-depth study of the performance and reliability aspects of dependable fault-tolerant systems. It identifies the current challenges that designers and practitioners must face in order to ensure the reliability, availability, and performance of systems, with special focus on their dynamic behaviors and dependencies, and provides system researchers, performance analysts, and practitioners with the tools to address these challenges in their work. With contributions from Prof. Trivedi's former PhD students and collaborators, many of whom are internationally recognized experts, to honor him on the occasion of his 70th birthday, this book serves as a valuable resource for all engineering disciplines, including electrical, computer, civil, mechanical, and industrial engineering as well as production and manufacturing.
This book is a concise navigator across the history of cybernetics, its state-of-the-art and prospects. The evolution of cybernetics (from N. Wiener to the present day) and the reasons of its ups and downs are presented. The correlation of cybernetics with the philosophy and methodology of control, as well as with system theory and systems analysis is clearly demonstrated. The book presents a detailed analysis focusing on the modern trends of research in cybernetics. A new development stage of cybernetics (the so-called cybernetics 2.0) is discussed as a science on general regularities of systems organization and control. The author substantiates the topicality of elaborating a new branch of cybernetics, i.e. organization theory which studies an organization as a property, process and system. The book is intended for theoreticians and practitioners, as well as for students, postgraduates and doctoral candidates. In the first place, the target audience includes tutors and lecturers preparing courses on cybernetics, control theory and systems science.
Cloud service benchmarking can provide important, sometimes surprising insights into the quality of services and leads to a more quality-driven design and engineering of complex software architectures that use such services. Starting with a broad introduction to the field, this book guides readers step-by-step through the process of designing, implementing and executing a cloud service benchmark, as well as understanding and dealing with its results. It covers all aspects of cloud service benchmarking, i.e., both benchmarking the cloud and benchmarking in the cloud, at a basic level. The book is divided into five parts: Part I discusses what cloud benchmarking is, provides an overview of cloud services and their key properties, and describes the notion of a cloud system and cloud-service quality. It also addresses the benchmarking lifecycle and the motivations behind running benchmarks in particular phases of an application lifecycle. Part II then focuses on benchmark design by discussing key objectives (e.g., repeatability, fairness, or understandability) and defining metrics and measurement methods, and by giving advice on developing own measurement methods and metrics. Next, Part III explores benchmark execution and implementation challenges and objectives as well as aspects like runtime monitoring and result collection. Subsequently, Part IV addresses benchmark results, covering topics such as an abstract process for turning data into insights, data preprocessing, and basic data analysis methods. Lastly, Part V concludes the book with a summary, suggestions for further reading and pointers to benchmarking tools available on the Web. The book is intended for researchers and graduate students of computer science and related subjects looking for an introduction to benchmarking cloud services, but also for industry practitioners who are interested in evaluating the quality of cloud services or who want to assess key qualities of their own implementations through cloud-based experiments.
Here's the first focused discussion of issues and technology in developing networked multimedia systems. This book includes a unique explanation of color specification and its role in achieving high picture quality, high compression ratio and high information retrieval performance, plus valuable coverage of principles and techniques of multimedia information indexing and retrieval critical for future systems.
Marking the 30th anniversary of the European Conference on Modelling and Simulation (ECMS), this inspirational text/reference reviews significant advances in the field of modelling and simulation, as well as key applications of simulation in other disciplines. The broad-ranging volume presents contributions from a varied selection of distinguished experts chosen from high-impact keynote speakers and best paper winners from the conference, including a Nobel Prize recipient, and the first president of the European Council for Modelling and Simulation (also abbreviated to ECMS). This authoritative book will be of great value to all researchers working in the field of modelling and simulation, in addition to scientists from other disciplines who make use of modelling and simulation approaches in their work. |
You may like...
Wireless Web Development with PHP and…
Ray Rischpater
Mixed media product
An Introduction to the Technique of…
Domenico Cantone, Pietro Ursino
Hardcover
R1,402
Discovery Miles 14 020
Algorithms as a Basis of Modern Applied…
Sarka Hoskova-Mayerova, Cristina Flaut, …
Hardcover
R5,232
Discovery Miles 52 320
Creating Business Applications with…
Jeffrey M Rhodes
Paperback
|