Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > General theory of computing > Systems analysis & design
Computing power performance was important at times when hardware was still expensive, because hardware had to be put to the best use. Later on this criterion was no longer critical, since hardware had become inexpensive. Meanwhile, however, people have realized that performance again plays a significant role, because of the major drain on system resources involved in developing complex applications. This book distinguishes between three levels of performance optimization: the system level, application level and business processes level. On each, optimizations can be achieved and cost-cutting potentials can be identified. The book presents the relevant theoretical background and measuring methods as well as proposed solutions. An evaluation of network monitors and checklists rounds out the work.
Systems for Online Transaction Processing (OLTP) and Online Analytical Processing (OLAP) are currently separate. The potential of the latest technologies and changes in operational and analytical applications over the last decade have given rise to the unification of these systems, which can be of benefit for both workloads. Research and industry have reacted and prototypes of hybrid database systems are now appearing. Benchmarks are the standard method for evaluating, comparing and supporting the development of new database systems. Because of the separation of OLTP and OLAP systems, existing benchmarks are only focused on one or the other. With the rise of hybrid database systems, benchmarks to assess these systems will be needed as well. Based on the examination of existing benchmarks, a new benchmark for hybrid database systems is introduced in this book. It is furthermore used to determine the effect of adding OLAP to an OLTP workload and is applied to analyze the impact of typically used optimizations in the historically separate OLTP and OLAP domains in mixed-workload scenarios.
This book presents intuitive explanations of the principles and applications of power system resiliency, as well as a number of straightforward and practical methods for the impact analysis of risk events on power system operations. It also describes the challenges of modelling, distribution networks, optimal scheduling, multi-stage planning, deliberate attacks, cyber-physical systems and SCADA-based smart grids, and how to overcome these challenges. Further, it highlights the resiliency issues using various methods, including strengthening the system against high impact events with low frequency and the fast recovery of the system properties. A large number of specialists have collaborated to provide innovative solutions and research in power systems resiliency. They discuss the fundamentals and contemporary materials of power systems resiliency, theoretical and practical issues, as well as current issues and methods for controlling the risk attacks and other threats to AC power systems. The book includes theoretical research, significant results, case studies, and practical implementation processes to offer insights into electric power and engineering and energy systems. Showing how systems should respond in case of malicious attacks, and helping readers to decide on the best approaches, this book is essential reading for electrical engineers, researchers and specialists. The book is also useful as a reference for undergraduate and graduate students studying the resiliency and reliability of power systems.
Formal methods are mathematically-based techniques, often supported
by reasoning tools, that can offer a rigorous and effective way to
model, design and analyze computer systems. The purpose of this
study is to evaluate international industrial experience in using
formal methods. The cases selected are representative of
industrial-grade projects and span a variety of application
domains. The study had three main objectives: - To better inform
deliberations within industry and government on standards and
regulations; - To provide an authoritative record on the practical
experience of formal methods to date; and
The demand for large-scale dependable, systems, such as Air Traffic Management, industrial plants and space systems, is attracting efforts of many word-leading European companies and SMEs in the area, and is expected to increase in the near future. The adoption of Off-The-Shelf (OTS) items plays a key role in such a scenario. OTS items allow mastering complexity and reducing costs and time-to-market; however, achieving these goals by ensuring dependability requirements at the same time is challenging. CRITICAL STEP project establishes a strategic collaboration between academic and industrial partners, and proposes a framework to support the development of dependable, OTS-based, critical systems. The book introduces methods and tools adopted by the critical systems industry, and surveys key achievements of the CRITICAL STEP project along four directions: fault injection tools, V&V of critical systems, runtime monitoring and evaluation techniques, and security assessment.
Fundamental Problems in Computing is in honor of Professor Daniel J. Rosenkrantz, a distinguished researcher in Computer Science. Professor Rosenkrantz has made seminal contributions to many subareas of Computer Science including formal languages and compilers, automata theory, algorithms, database systems, very large scale integrated systems, fault-tolerant computing and discrete dynamical systems. For many years, Professor Rosenkrantz served as the Editor-in-Chief of the Journal of the Association for Computing Machinery (JACM), a very prestigious archival journal in Computer Science. His contributions to Computer Science have earned him many awards including the Fellowship from ACM and the ACM SIGMOD Contributions Award.
With the rise of mobile and wireless technologies, more sustainable networks are necessary to support communication. These next-generation networks can now be utilized to extend the growing era of the Internet of Things. Enabling Technologies and Architectures for Next-Generation Networking Capabilities is an essential reference source that explores the latest research and trends in large-scale 5G technologies deployment, software-defined networking, and other emerging network technologies. Featuring research on topics such as data management, heterogeneous networks, and spectrum sensing, this book is ideally designed for computer engineers, technology developers, network administrators and researchers, professionals, and graduate-level students seeking coverage on current and future network technologies.
The creation and consumption of content, especially visual content, is ingrained into our modern world. This book contains a collection of texts centered on the evaluation of image retrieval systems. To enable reproducible evaluation we must create standardized benchmarks and evaluation methodologies. The individual chapters in this book highlight major issues and challenges in evaluating image retrieval systems and describe various initiatives that provide researchers with the necessary evaluation resources. In particular they describe activities within ImageCLEF, an initiative to evaluate cross-language image retrieval systems which has been running as part of the Cross Language Evaluation Forum (CLEF) since 2003. To this end, the editors collected contributions from a range of people: those involved directly with ImageCLEF, such as the organizers of specific image retrieval or annotation tasks; participants who have developed techniques to tackle the challenges set forth by the organizers; and people from industry and academia involved with image retrieval and evaluation generally. Mostly written for researchers in academia and industry, the book stresses the importance of combing textual and visual information - a multimodal approach - for effective retrieval. It provides the reader with clear ideas about information retrieval and its evaluation in contexts and domains such as healthcare, robot vision, press photography, and the Web.
SystemVerilog is a rich set of extensions to the IEEE 1364-2001 Verilog Hardware Description Language (Verilog HDL). These extensions address two major aspects of HDL-based design. First, modeling very large designs with concise, accurate, and intuitive code. Second, writing high-level test programs to efficiently and effectively verify these large designs. The first edition of this book addressed the first aspect of the SystemVerilog extensions to Verilog. Important modeling features were presented, such as two-state data types, enumerated types, user-degined types, structures, unions, and interfaces. Emphasis was placed on the proper usage of these enhancements for simulation and synthesis.
"Kent is a master at creating code that communicates well, is easy to understand, and is a pleasure to read. Every chapter of this book contains excellent explanations and insights into the smaller but important decisions we continuously have to make when creating quality code and classes." -Erich Gamma, IBM Distinguished Engineer "Many teams have a master developer who makes a rapid stream of good decisions all day long. Their code is easy to understand, quick to modify, and feels safe and comfortable to work with. If you ask how they thought to write something the way they did, they always have a good reason. This book will help you become the master developer on your team. The breadth and depth of topics will engage veteran programmers, who will pick up new tricks and improve on old habits, while the clarity makes it accessible to even novice developers." -Russ Rufer, Silicon Valley Patterns Group "Many people don't realize how readable code can be and how valuable that readability is. Kent has taught me so much, I'm glad this book gives everyone the chance to learn from him." -Martin Fowler, chief scientist, ThoughtWorks "Code should be worth reading, not just by the compiler, but by humans. Kent Beck distilled his experience into a cohesive collection of implementation patterns. These nuggets of advice will make your code truly worth reading." -Gregor Hohpe, author of Enterprise Integration Patterns "In this book Kent Beck shows how writing clear and readable code follows from the application of simple principles. Implementation Patterns will help developers write intention revealing code that is both easy to understand and flexible towards future extensions. A must read for developers who are serious about their code." -Sven Gorts "Implementation Patterns bridges the gap between design and coding. Beck introduces a new way of thinking about programming by basing his discussion on values and principles." -Diomidis Spinellis, author of Code Reading and Code Quality Software Expert Kent Beck Presents a Catalog of Patterns Infinitely Useful for Everyday Programming Great code doesn't just function: it clearly and consistently communicates your intentions, allowing other programmers to understand your code, rely on it, and modify it with confidence. But great code doesn't just happen. It is the outcome of hundreds of small but critical decisions programmers make every single day. Now, legendary software innovator Kent Beck-known worldwide for creating Extreme Programming and pioneering software patterns and test-driven development-focuses on these critical decisions, unearthing powerful "implementation patterns" for writing programs that are simpler, clearer, better organized, and more cost effective. Beck collects 77 patterns for handling everyday programming tasks and writing more readable code. This new collection of patterns addresses many aspects of development, including class, state, behavior, method, collections, frameworks, and more. He uses diagrams, stories, examples, and essays to engage the reader as he illuminates the patterns. You'll find proven solutions for handling everything from naming variables to checking exceptions. This book covers The value of communicating through code and the philosophy behind patterns How and when to create classes, and how classes encode logic Best practices for storing and retrieving state Behavior: patterns for representing logic, including alternative paths Writing, naming, and decomposing methods Choosing and using collections Implementation pattern variations for use in building frameworks Implementation Patterns will help programmers at all experience levels, especially those who have benefited from software patterns or agile methods. It will also be an indispensable resource for development teams seeking to work together more efficiently and build more maintainable software. No other programming book will touch your day-to-day work more often.
The 2008 TUB-SJTU joint workshop on Autonomous Systems Self-Organization, Management, and Control was held on October 6, 2008 at Shanghai Jiao Tong University, Shanghai, China. The workshop, sponsored by Shanghai Jiao Tong University and Technical University of Berlin brought together scientists and researchers from both universities to present and discuss the latest progress on autonomous systems and its applications in diverse areas. Autonomous systems are designed to integrate machines, computing, sensing, and software to create intelligent systems capable of interacting with the complexities of the real world. Autonomous systems represent the physical embodiment of machine intelligence. Topics of interest include, but are not limited to theory and modeling for autonomous systems; organization of autonomous systems; learning and perception; complex systems; multi-agent systems; robotics and control; applications of autonomous systems.
This book presents the technical program of the International Embedded Systems Symposium (IESS) 2009. Timely topics, techniques and trends in embedded system design are covered by the chapters in this volume, including modelling, simulation, verification, test, scheduling, platforms and processors. Particular emphasis is paid to automotive systems and wireless sensor networks. Sets of actual case studies in the area of embedded system design are also included. Over recent years, embedded systems have gained an enormous amount of proce- ing power and functionality and now enter numerous application areas, due to the fact that many of the formerly external components can now be integrated into a single System-on-Chip. This tendency has resulted in a dramatic reduction in the size and cost of embedded systems. As a unique technology, the design of embedded systems is an essential element of many innovations. Embedded systems meet their performance goals, including real-time constraints, through a combination of special-purpose hardware and software components tailored to the system requirements. Both the development of new features and the reuse of existing intellectual property components are essential to keeping up with ever more demanding customer requirements. Furthermore, design complexities are steadily growing with an increasing number of components that have to cooperate properly. Embedded system designers have to cope with multiple goals and constraints simul- neously, including timing, power, reliability, dependability, maintenance, packaging and, last but not least, price.
This book aims to deconstruct ethnography to alert systems designers, and other stakeholders, to the issues presented by new approaches that move beyond the studies of 'work' and 'work practice' within the social sciences (in particular anthropology and sociology). The theoretical and methodological apparatus of the social sciences distort the social and cultural world as lived in and understood by ordinary members, whose common-sense understandings shape the actual milieu into which systems are placed and used. In Deconstructing Ethnography the authors show how 'new' calls are returning systems design to 'old' and problematic ways of understanding the social. They argue that systems design can be appropriately grounded in the social through the ordinary methods that members use to order their actions and interactions. This work is written for post-graduate students and researchers alike, as well as design practitioners who have an interest in bringing the social to bear on design in a systematic rather than a piecemeal way. This is not a 'how to' book, but instead elaborates the foundations upon which the social can be systematically built into the design of ubiquitous and interactive systems.
Requirements Management has proven itself to be an enormous potential for the optimization of development projects throughout the last few years. Especially in the climate of an increasingly competitive market Requirements Management helps in carrying out developments faster, cheaper and with a higher quality. This book focuses on the interfaces of Requirements Management to the other disciplines of Systems Engineering, for example Project Management, Change Management and Configuration and Version Management. To this end, an introduction into Requirements Management and Requirements Development is given, along with a short sketch of Systems Engineering, and especially the necessary inputs and resulting outputs of Requirements Management are explained. Using these flows of information it is shown how Requirements Management can support and optimize the other project disciplines and how very important therefore a functioning Requirements Management is for all areas of development.
This book is dedicated to Prof. Dr. Heinz Gerhauser on the occasion of his retirement both from the position of Executive Director of the Fraunhofer Institute for Integrated Circuits IIS and from the Endowed Chair of Information Technologies with a Focus on Communication Electronics (LIKE) at the Friedrich-Alexander-Universitat Erlangen-Nurnberg. Heinz Gerhauser's vision and entrepreneurial spirit have made the Fraunhofer IIS one of the most successful and renowned German research institutions. He has been Director of the Fraunhofer IIS since 1993, and under his leadership it has grown to become the largest of Germany's 60 Fraunhofer Institutes, a position it retains to this day, currently employing over 730 staff. Likely his most important scientific as well as application-related contribution was his pivotal role in the development of the mp3 format, which would later become a worldwide success. The contributions to this Festschrift were written by both Fraunhofer IIS staff and external project team members in appreciation of Prof. Dr. Gerhauser's lifetime academic achievements and his inspiring leadership at the Fraunhofer IIS. The papers reflect the broad spectrum of the institute's research activities and are grouped into sections on circuits, information systems, visual computing, and audio and multimedia. They provide academic and industrial researchers in fields like signal processing, sensor networks, microelectronics, and integrated circuits with an up-to-date overview of research results that have a huge potential for cutting-edge industrial applications.
The most significant articles from each of the fields represented at the conference on Work with Display Units 1992 are presented in this volume. Such topics are:
This book is open access under a CC BY NC ND license. It addresses the most recent developments in cloud computing such as HPC in the Cloud, heterogeneous cloud, self-organising and self-management, and discusses the business implications of cloud computing adoption. Establishing the need for a new architecture for cloud computing, it discusses a novel cloud management and delivery architecture based on the principles of self-organisation and self-management. This focus shifts the deployment and optimisation effort from the consumer to the software stack running on the cloud infrastructure. It also outlines validation challenges and introduces a novel generalised extensible simulation framework to illustrate the effectiveness, performance and scalability of self-organising and self-managing delivery models on hyperscale cloud infrastructures. It concludes with a number of potential use cases for self-organising, self-managing clouds and the impact on those businesses.
This title uses an approach to systems analysis and modelling with a systems science flavour that stimulates systems thinking. The author moves from illustrating the three-fold nature of the universe to circumscribing a "Yellowstone National Park" grizzly bear system. After introducing systems modelling principles, the ensuing selecting of examples illustrate that anything which changes over time can be modelled as a system. Each example begins with a knowledge base whose items display relevant information obtained from systems analysis. Repetition of this approach for a diversity of examples clearly establishes a new protocol for synthesising systems models. New ideas such as dynamics forms and reverse regression are merged with the well-known continuity equation to synthesise linear-equation models of nonlinear and/or stochastic systems, thus replacing the more conventional differential equation models. Standard matrix algebra is employed in their solution.
This book presents a comprehensive introduction to Internetware, covering aspects ranging from the fundamental principles and engineering methodologies to operational platforms, quality measurements and assurance and future directions. It also includes guidelines and numerous representative real-world case studies that serve as an invaluable reference resource for software engineers involved in the development of Internetware applications. Providing a detailed analysis of current trends in modern software engineering in the Internet, it offers an essential blueprint and an important contribution to the research on software engineering and systems for future Internet computing.
This book focuses on the basic control and filtering synthesis problems for discrete-time switched linear systems under time-dependent switching signals. Chapter 1, as an introduction of the book, gives the backgrounds and motivations of switched systems, the definitions of the typical time-dependent switching signals, the differences and links to other types of systems with hybrid characteristics and a literature review mainly on the control and filtering for the underlying systems. By summarizing the multiple Lyapunov-like functions (MLFs) approach in which different requirements on comparisons of Lyapunov function values at switching instants, a series of methodologies are developed for the issues on stability and stabilization, and l2-gain performance or tube-based robustness for l disturbance, respectively, in Chapters 2 and 3. Chapters 4 and 5 are devoted to the control and filtering problems for the time-dependent switched linear systems with either polytopic uncertainties or measurable time-varying parameters in different sense of disturbances. The asynchronous switching problem, where there is time lag between the switching of the currently activated system mode and the controller/filter to be designed, is investigated in Chapter 6. The systems with various time delays under typical time-dependent switching signals are addressed in Chapter 7.
This book is open access under a CC BY-NC 2.5 license. This book presents the VISCERAL project benchmarks for analysis and retrieval of 3D medical images (CT and MRI) on a large scale, which used an innovative cloud-based evaluation approach where the image data were stored centrally on a cloud infrastructure and participants placed their programs in virtual machines on the cloud. The book presents the points of view of both the organizers of the VISCERAL benchmarks and the participants. The book is divided into five parts. Part I presents the cloud-based benchmarking and Evaluation-as-a-Service paradigm that the VISCERAL benchmarks used. Part II focuses on the datasets of medical images annotated with ground truth created in VISCERAL that continue to be available for research. It also covers the practical aspects of obtaining permission to use medical data and manually annotating 3D medical images efficiently and effectively. The VISCERAL benchmarks are described in Part III, including a presentation and analysis of metrics used in evaluation of medical image analysis and search. Lastly, Parts IV and V present reports by some of the participants in the VISCERAL benchmarks, with Part IV devoted to the anatomy benchmarks and Part V to the retrieval benchmark. This book has two main audiences: the datasets as well as the segmentation and retrieval results are of most interest to medical imaging researchers, while eScience and computational science experts benefit from the insights into using the Evaluation-as-a-Service paradigm for evaluation and benchmarking on huge amounts of data.
Increasing system complexity has created a pressing need for better design tools and associated methodologies and languages for meeting the stringent time to market and cost constraints. Platform-centric and platfo- based system-on-chip (SoC) design methodologies, based on reuse of software and hardware functionality, has also gained increasing exposure and usage within the Electronic System-Level (ESL) design communities. The book proposes a new methodology for realizing platform-centric design of complex systems, and presents a detailed plan for its implementation. The proposed plan allows component vendors, system integrators and product developers to collaborate effectively and efficiently to create complex products within budget and schedule constraints. This book focuses more on the use of platforms in the design of products, and not on the design of platforms themselves. Platform-centric design is not for everyone, as some may feel that it does not allow them to differentiate their offering from competitors to a significant degree. However, its proponents may claim that the time-- market and cost advantages of platform-centric design more than compensate for any drawbacks.
In 1998-99, at the dawn of the SoC Revolution, we wrote Surviving the SOC Revolution: A Guide to Platform Based Design. In that book, we focused on presenting guidelines and best practices to aid engineers beginning to design complex System-on-Chip devices (SoCs). Now, in 2003, facing the mid-point of that revolution, we believe that it is time to focus on winning. In this book, Winning the SoC Revolution: Experiences in Real Design, we gather the best practical experiences in how to design SoCs from the most advanced design groups, while setting the issues and techniques in the context of SoC design methodologies. As an edited volume, this book has contributions from the leading design houses who are winning in SoCs - Altera, ARM, IBM, Philips, TI, UC Berkeley, and Xilinx. These chapters present the many facets of SoC design - the platform based approach, how to best utilize IP, Verification, FPGA fabrics as an alternative to ASICs, and next generation process technology issues. We also include observations from Ron Wilson of CMP Media on best practices for SoC design team collaboration. We hope that by utilizing this book, you too, will win the SoC Revolution.
This volume provides an introduction to and overview of the emerging field of interconnected networks which include multilayer or multiplex networks, as well as networks of networks. Such networks present structural and dynamical features quite different from those observed in isolated networks. The presence of links between different networks or layers of a network typically alters the way such interconnected networks behave - understanding the role of interconnecting links is therefore a crucial step towards a more accurate description of real-world systems. While examples of such dissimilar properties are becoming more abundant - for example regarding diffusion, robustness and competition - the root of such differences remains to be elucidated. Each chapter in this topical collection is self-contained and can be read on its own, thus making it also suitable as reference for experienced researchers wishing to focus on a particular topic.
"Satellite Network Robust QoS-aware Routing" presents a novel routing strategy for satellite networks. This strategy is useful for the design of multi-layered satellite networks as it can greatly reduce the number of time slots in one system cycle. The traffic prediction and engineering approaches make the system robust so that the traffic spikes can be handled effectively. The multi-QoS optimization routing algorithm can satisfy various potential user requirements. Clear and sufficient illustrations are also presented in the book. As the chapters cover the above topics independently, readers from different research backgrounds in constellation design, multi-QoS routing, and traffic engineering can benefit from the book. Fei Long is a senior engineer at Beijing R&D Center of 54th Research Institute of China Electronics Technology Group Corporation. |
You may like...
Handbook of Research on 5G Networks and…
Augustine O Nwajana, Isibor Kennedy Ihianle
Hardcover
R8,415
Discovery Miles 84 150
Information Systems, International…
Ralph Stair, George Reynolds
Paperback
Implementing Data Analytics and…
Chintan Bhatt, Neeraj Kumar, …
Hardcover
R6,256
Discovery Miles 62 560
Cases on Lean Thinking Applications in…
Eduardo Guilherme Satolo, Robisom Damasceno Calado
Hardcover
R6,281
Discovery Miles 62 810
Systems Analysis And Design In A…
John Satzinger, Robert Jackson, …
Hardcover
(1)
|