![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > General theory of computing > Systems analysis & design
Learn to solve the unprecedented challenges facing Online Learning and Adaptive Signal Processing in this concise, intuitive text. The ever-increasing amount of data generated every day requires new strategies to tackle issues such as: combining data from a large number of sensors; improving spectral usage, utilizing multiple-antennas with adaptive capabilities; or learning from signals placed on graphs, generating unstructured data. Solutions to all of these and more are described in a condensed and unified way, enabling you to expose valuable information from data and signals in a fast and economical way. The up-to-date techniques explained here can be implemented in simple electronic hardware, or as part of multi-purpose systems. Also featuring alternative explanations for online learning, including newly developed methods and data selection, and several easily implemented algorithms, this one-of-a-kind book is an ideal resource for graduate students, researchers, and professionals in online learning and adaptive filtering.
The book is directly relevant for students on HND, degree and professional courses. This market-leading text covers the whole range of activities necessary for the analysis, design and implementation of computer-based information and data processing systems. The authors emphasize the role of people, management and quality issues, and consider practical and business realities.
The measurement of dependability attributes on real systems is a very time-consuming and costly affair, making analytical or simulation modeling the only viable solutions. Dependability of Networked Computer-based Systems explores reliability, availability and safety modeling of networked computer-based systems used in life-critical applications such as avionics, nuclear power plants, automobiles and chemical process industries. Dependability of Networked Computer-based Systems gives an overview of basic dependability modeling concepts and addresses new challenges in dependability modeling of networked computer-based systems, as well as new trends, their capabilities and limitations. It covers a variety of dependability modeling methods: stochastic processes, Markov and semi-Markov models, response-time distribution, stochastic Petri-net-based modeling formalisms, and Monte Carlo simulation models. Dependability of Networked Computer-based Systems provides students and researchers with a detailed overview of dependability models and analysis techniques. Practicing engineers will also find this text a useful guide to decision-making based on system dependability at the design, operation and maintenance stages.
This introductory user's guide to systems analysis and systems design focuses on building sustainable information systems to meet tomorrow's needs. It shows how practitioners can apply multiple participatory perspectives in development, so as to avoid future problems. As a practical guide, it is presented to be readily comprehensible and is organized to enable users to concentrate on their goals efficiently, and with minimum theoretical elaboration. The chapters follow the sequence involved in planning an information system, explaining key words, the time involved in each step, ending with a tutorial or exercises.
This introductory user's guide to systems analysis and systems design focuses on building sustainable information systems to meet tomorrow's needs. It shows how practitioners can apply multiple participatory perspectives in development, so as to avoid future problems. As a practical guide, it is presented to be readily comprehensible and is organized to enable users to concentrate on their goals efficiently, and with minimum theoretical elaboration. The chapters follow the sequence involved in planning an information system, explaining key words, the time involved in each step, ending with a tutorial or exercises.
Prototyping and user testing is the best way to create successful products, but many designers skip this important step and use gut instinct instead. By explaining the goals and methodologies behind prototyping-and demonstrating how to prototype for both physical and digital products-this practical guide helps beginning and intermediate designers become more comfortable with creating and testing prototypes early and often in the process. Author Kathryn McElroy explains various prototyping methods, from fast and dirty to high fidelity and refined, and reveals ways to test your prototypes with users. You'll gain valuable insights for improving your product, whether it's a smartphone app or a new electronic gadget. Learn similarities and differences between prototyping for physical and digital products Know what fidelity level is needed for different prototypes Get best practices for prototyping in a variety of mediums, and choose which prototyping software or components to use Learn electronics prototyping basics and resources for getting started Write basic pseudocode and translate it into usable code for Arduino Conduct user tests to gain insights from prototypes
The control-flow issues presented in this textbook are extremely relevant in modern computer languages and programming styles. In addition to the basic control-flow mechanisms, virtually all new computer languages provide some form of exceptional control flow to support robust programming introduced in this textbook. Also, concurrency capabilities are appearing with increasing frequency in both new and old programming languages, and are covered in this book. Understanding Control Flow: With Concurrent Programming Using C++ starts with looping, and works through each of the basic control-flow concepts, examining why each is fundamental and where it is useful. Time is spent on each concept according to its level of difficulty. Examples and exercises are also provided in this textbook. New programming methodologies are requiring new forms of control flow, and new programming languages are supporting these methodologies with new control structures, such as the concurrency constructs discussed in this textbook. Most computers now contain multi-threading and multi-cores, while multiple processors and distributed systems are ubiquitous - all of which require advanced programming methodologies to take full advantage of the available parallelism summarized in this textbook. Advance forms of control flow are becoming basic programming skills needed by all programmers, not just graduate students working in the operating systems or database disciplines. This textbook is designed for advanced-level students studying computer science and engineering. Professionals and researchers working in this field, specifically programming and software engineering, will find this book useful as a reference.
"Details the most recent advances in Laboratory Information Management Systems. Offers contemporary approaches to system development, design, and installation; system customization; software and hardware compatibility; quality assurance and regulatory requirements; and resource utilization."
A comprehensive guide to understanding the standard and most recent advances in the design of reliable computer systems. It is organized into three sections, beginning with an in-depth review of existing reliability techniques and evaluation criteria for both hardware and software. also examined are the models for detecting faults and predicting failures, and the financial considerations which are inherent in the design, purchase, operation, and maintenance of a reliable system.
This book investigates the fundaments, standards, and protocols of Cyber-Physical Systems (CPS) in the industrial transformation environment. It facilitates a fusion of both technologies in the creation of reliable and robust applications. Cyber-Physical Systems for Industrial Transformation: Fundamentals, Standards, and Protocols explores emerging technologies such as artificial intelligence, data science, blockchain, robotic process automation, virtual reality, edge computing, and 5G technology to highlight current and future opportunities to transition CPS to become more robust and reliable. The book showcases the real-time sensing, processing, and actuation software and discusses fault-tolerant and cybersecurity as well. This book brings together undergraduates, postgraduates, academics, researchers, and industry individuals that are interested in exploring new ideas, techniques, and tools related to CPS and Industry 4.0.
Embedded computer systems literally surround us: they're in our cell phones, PDAs, cars, TVs, refrigerators, heating systems, and more. In fact, embedded systems are one of the most rapidly growing segments of the computer industry today. Along with the growing list of devices for which embedded computer systems are appropriate, interest is growing among programmers, hobbyists, and engineers of all types in how to design and build devices of their own. Furthermore, the knowledge offered by this book into the fundamentals of these computer systems can benefit anyone who has to evaluate and apply the systems. The second edition of "Designing Embedded Hardware" has been updated to include information on the latest generation of processors and microcontrollers, including the new MAXQ processor. If you're new to this and don't know what a MAXQ is, don't worry--the book spells out the basics of embedded design for beginners while providing material useful for advanced systems designers. "Designing Embedded Hardware" steers a course between those books dedicated to writing code for particular microprocessors, and those that stress the philosophy of embedded system design without providing any practical information. Having designed 40 embedded computer systems of his own, author John Catsoulis brings a wealth of real-world experience to show readers how to design and create entirely new embedded devices and computerized gadgets, as well as how to customize and extend off-the-shelf systems. Loaded with real examples, this book also provides a roadmap to the pitfalls and traps to avoid. "Designing Embedded Hardware" includes: The theory and practice of embedded systems Understanding schematics and data sheets Powering an embedded system Producing and debugging an embedded system Processors such as the PIC, Atmel AVR, and Motorola 68000-series Digital Signal Processing (DSP) architectures Protocols (SPI and I2C) used to add peripherals RS-232C, RS-422, infrared communication, and USB CAN and Ethernet networking Pulse Width Monitoring and motor control If you want to build your own embedded system, or tweak an existing one, this invaluable book gives you the understanding and practical skills you need.
This book presents state-of-the-art solutions to the theoretical and practical challenges stemming from the leverage of big data and its computational intelligence in supporting smart network operation, management, and optimization. In particular, the technical focus covers the comprehensive understanding of network big data, efficient collection and management of network big data, distributed and scalable online analytics for network big data, and emerging applications of network big data for computational intelligence.
Codesign for Real-Time Video Applications describes a modern design approach for embedded systems. It combines the design of hardware, software, and algorithms. Traditionally, these design domains are treated separately to reduce the design complexity. Advanced design tools support a codesign of the different domains which opens an opportunity for exploiting synergetic effects. The design approach is illustrated by the design of a video compression system. It is integrated into the video card of a PC. A VLIW processor architecture is used as the basis of the compression system and popular video compression algorithms (MPEG, JPEG, H.261) are analyzed. A complete top-down design flow is presented and the design tools for each of the design steps are explained. The tools are integrated into an HTML-based design framework. The resulting design data can be directly integrated into the WWW. This is a crucial aspect for supporting distributed design groups. The design data can be directly documented an cross referencing in an almost arbitrary way is supported. This provides a platform for information sharing among the different design domains. Codesign for Real-Time Video Applications focuses on the multi-disciplinary aspects of embedded system design. It combines design automation and advanced processor design with an important application domain. A quantitative design approach is emphasized which focuses the design time on the most crucial components. Thus enabling a fast and cost efficient design methodology. This book will be of interest to researchers, designers and managers working in embedded system design.
The next generation of computer system designers will be less concerned about details of processors and memories, and more concerned about the elements of a system tailored to particular applications. These designers will have a fundamental knowledge of processors and other elements in the system, but the success of their design will depend on the skills in making system-level tradeoffs that optimize the cost, performance and other attributes to meet application requirements. This book provides a new treatment of computer system design, particularly for System-on-Chip (SOC), which addresses the issues mentioned above. It begins with a global introduction, from the high-level view to the lowest common denominator (the chip itself), then moves on to the three main building blocks of an SOC (processor, memory, and interconnect). Next is an overview of what makes SOC unique (its customization ability and the applications that drive it). The final chapter presents future challenges for system design and SOC possibilities.
As the uses of digital systems continue to proliferate in quantity and variety, field programmable gate arrays (FPGAs) are taking centre-stage in their design. Introduction to Embedded System Design Using Field Programmable Gate Arrays provides a starting point for the use of FPGAs in the design of embedded systems. The text considers a hypothetical robot controller as an embedded application and weaves around it related concepts of FPGA-based digital design. The book details: a [ use of FPGA vis-A -vis general purpose processor and microcontroller; a [ design using Verilog hardware description language; a [ digital design synthesis using Verilog and XilinxA(R) SpartanTM 3 FPGA; a [ FPGA-based embedded processors and peripherals; a [ overview of serial data communications and signal conditioning using FPGA; a [ FPGA-based motor drive controllers; and a [ prototyping digital systems using FPGA. The book is a good introductory text for FPGA-based design. It will be useful for both students and designers who have worked with microprocessors and microcontrollers and now wish to move to design using FPGA. Its end-of-chapter exercises and frequent use of example can be used for teaching or for self-study.
This book focuses on futuristic approaches and designs for real-time systems and applications, as well as the fundamental concepts of including advanced techniques and tools in models of data-driven blockchain ecosystems. The Data-Driven Blockchain Ecosystem: Fundamentals, Applications, and Emerging Technologies discusses how to implement and manage processes for releasing and delivering blockchain applications. It presents the core of blockchain technology, IoT-based and AI-based blockchain systems, and various manufacturing areas related to Industry 4.0. The book illustrates how to apply design principles to develop and manage blockchain networks, and also covers the role that cloud computing plays in blockchain applications. All major technologies involved in blockchain-embedded applications are included in this book, which makes it useful to engineering students, researchers, academicians, and professionals interested in the core of blockchain technology.
Research 2.0 and the Future of Information Literacy examines possible congruencies between information literacy and Research 2.0, because the work of today's researcher mobilizes a number of literacies. From among the various types of relevant literacies, at least three types of literacies can be mentioned in this relation: information literacy, scientific literacy and academic literacy. This book addresses these literacies in the light of the changing research landscape. Broad contexts of the researcher's abilities, as adaptive and innovative thinking, problem solving skills, self-management and design mindset are also examined. Computational thinking and the computational paradigm in a number of fields of research are taken into consideration, as well. Researchers differ to non-researchers when populating social media, which means that these two different groups require different literacies. The relationship between information literacy and information is approached in a new way. Among the multitude of issues, we introduce a new interface between information literacy and Research 2.0. It encompasses the issues of research data management and data literacy, which represent also a challenge both for the academic library and for the communities of researchers. Similarly, the questions of new metrics of scientific output are addressed in the book.
Research into grid computing has been driven by the need to solve large-scale, increasingly complex problems for scientific applications. Yet the applications of grid computing for business and casual users did not begin to emerge until the development of the concept of cloud computing, fueled by advances in virtualization techniques, coupled with the increased availability of ever-greater Internet bandwidth. The appeal of this new paradigm is mainly based on its simplicity, and the affordable price for seamless access to both computational and storage resources. This timely text/reference introduces the fundamental principles and techniques underlying grids, clouds and virtualization technologies, as well as reviewing the latest research and expected future developments in the field. Readers are guided through the key topics by internationally recognized experts, enabling them to develop their understanding of an area likely to play an ever more significant role in coming years. Topics and features: presents contributions from an international selection of experts in the field; provides a thorough introduction and overview of existing technologies in grids, clouds and virtualization, including a brief history of the field; examines the basic requirements for performance isolation of virtual machines on multi-core servers, analyzing a selection of system virtualization technologies; examines both business and scientific applications of grids and clouds, including their use in the life sciences and for high-performance computing; explores cloud building technologies, architectures for enhancing grid infrastructures with cloud computing, and cloud performance; discusses energy aware grids and clouds, workflows on grids and clouds, and cloud and grid programming models. This useful text will enable interested readers to familiarize themselves with the key topics of grids, clouds and virtualization, and to contribute to new advances in the field. Researchers, undergraduate and graduate students, system designers and programmers, and IT policy makers will all benefit from the material covered.
Transition Engineering: Building a Sustainable Future examines new strategies emerging in response to the mega-issues of global climate change, decline in world oil supply, scarcity of key industrial minerals, and local environmental constraints. These issues pose challenges for organizations, businesses, and communities, and engineers will need to begin developing ideas and projects to implement the transition of engineered systems. This work presents a methodology for shifting away from unsustainable activities. Teaching the Transition Engineering approach and methodology is the focus of the text, and the concept is presented in a way that engineers can begin applying it in their work.
Learn how to use, deploy, and maintain Apache Spark with this comprehensive guide, written by the creators of the open-source cluster-computing framework. With an emphasis on improvements and new features in Spark 2.0, authors Bill Chambers and Matei Zaharia break down Spark topics into distinct sections, each with unique goals. You'll explore the basic operations and common functions of Spark's structured APIs, as well as Structured Streaming, a new high-level API for building end-to-end streaming applications. Developers and system administrators will learn the fundamentals of monitoring, tuning, and debugging Spark, and explore machine learning techniques and scenarios for employing MLlib, Spark's scalable machine-learning library. Get a gentle overview of big data and Spark Learn about DataFrames, SQL, and Datasets-Spark's core APIs-through worked examples Dive into Spark's low-level APIs, RDDs, and execution of SQL and DataFrames Understand how Spark runs on a cluster Debug, monitor, and tune Spark clusters and applications Learn the power of Structured Streaming, Spark's stream-processing engine Learn how you can apply MLlib to a variety of problems, including classification or recommendation
Software components and component-based software development (CBSD) are acknowledged as the best approach for constructing quality software at reasonable cost. Composing Software Components: A Software-testing Perspective describes a 10-year investigation into the underlying principles of CBSD. By restricting attention to the simplest cases, startling results are obtained: Components are tested using only executable code. Their behavior is recorded and presented graphically. Functional and non-functional behavior of systems synthesized from components are calculated from component tests alone. No access to components themselves is required. Fast, accurate tools support every aspect of CBSD from design through debugging. Case studies of CBSD also illuminate software testing in general, particularly an expanded role for unit testing and the treatment of non-functional software properties. This unique book: Contains more than a dozen case studies of fully worked-out component synthesis, with revealing insights into fundamental testing issues. Presents an original, fundamental theory of component composition that includes persistent state and concurrency, based on functional software testing rather than proof-of-programs. Comes with free supporting software with tutorial examples and data for replication of examples. The Perl software has been tested on Linux, Macintosh, and Windows platforms. Full documentation is provided. Includes anecdotes and insights from the author's 50-year career in computing as systems programmer, manager, researcher, and teacher. Composing Software Components: A Software-testing Perspective will help software researchers and practitioners to understand the underlying principles of component testing. Advanced students in computer science, engineering, and mathematics can also benefit from the book as a supplemental text and reference."
The present work showcases a novel approach to modeling systems architectures by utilizing Lego bricks and RFID technology. The presented solution can be used by systems and software architects to communicate their design decisions with other stakeholders in the developments process such as customers and managers involved. The software provided in this book helps to get a concrete tool showing how the approach can be applied. If the reader is interested in experimenting with this approach, they will need to purchase LEGO (c) blocks and the required RFID technology needed for this.
The Intelligent Systems Series comprises titles that present state-of-the-art knowledge and the latest advances in intelligent systems. Its scope includes theoretical studies, design methods, and real-world implementations and applications. The most prevalent topics in Intelligence and Security
Informatics (ISI) include data management, data and text mining for
ISI applications, terrorism informatics, deception and intent
detection, terrorist and criminal social network analysis, public
health and bio-security, crime analysis, cyber-infrastructure
protection, transportation infrastructure security, policy studies
and evaluation, and information assurance, among others.This book
covers the most active research work inrecent years.
This updated edition offers an indispensable exposition on real-time computing, with particular emphasis on predictable scheduling algorithms. It introduces the fundamental concepts of real-time computing, demonstrates the most significant results in the field, and provides the essential methodologies for designing predictable computing systems used to support time-critical control applications. Along with an in-depth guide to the available approaches for the implementation and analysis of real-time applications, this revised edition contains a close examination of recent developments in real-time systems, including limited preemptive scheduling, resource reservation techniques, overload handling algorithms, and adaptive scheduling techniques. This volume serves as a fundamental advanced-level textbook. Each chapter provides basic concepts, which are followed by algorithms, illustrated with concrete examples, figures and tables. Exercises and solutions are provided to enhance self-study, making this an excellent reference for those interested in real-time computing for designing and/or developing predictable control applications. |
You may like...
Dynamic Auditing - A Student Edition
B. Marx, A. van der Watt, …
Paperback
Death And Taxes - How SARS Made Hitmen…
Johann van Loggerenberg
Paperback
Biographical Dictionary of the Board of…
Bernard S. Katz
Hardcover
Global Capital and National Politics…
Timothy Kessler
Hardcover
|