![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > General theory of computing > Systems analysis & design
The two volumes LNCS 8805 and 8806 constitute the thoroughly refereed post-conference proceedings of 18 workshops held at the 20th International Conference on Parallel Computing, Euro-Par 2014, in Porto, Portugal, in August 2014. The 100 revised full papers presented were carefully reviewed and selected from 173 submissions. The volumes include papers from the following workshops: APCI&E (First Workshop on Applications of Parallel Computation in Industry and Engineering - BigDataCloud (Third Workshop on Big Data Management in Clouds) - DIHC (Second Workshop on Dependability and Interoperability in Heterogeneous Clouds) - FedICI (Second Workshop on Federative and Interoperable Cloud Infrastructures) - Hetero Par (12th International Workshop on Algorithms, Models and Tools for Parallel Computing on Heterogeneous Platforms) - HiBB (5th Workshop on High Performance Bioinformatics and Biomedicine) - LSDVE (Second Workshop on Large Scale Distributed Virtual Environments on Clouds and P2P) - MuCoCoS (7th International Workshop on Multi-/Many-core Computing Systems) - OMHI (Third Workshop on On-chip Memory Hierarchies and Interconnects) - PADAPS (Second Workshop on Parallel and Distributed Agent-Based Simulations) - PROPER (7th Workshop on Productivity and Performance) - Resilience (7th Workshop on Resiliency in High Performance Computing with Clusters, Clouds, and Grids) - REPPAR (First International Workshop on Reproducibility in Parallel Computing) - ROME (Second Workshop on Runtime and Operating Systems for the Many Core Era) - SPPEXA (Workshop on Software for Exascale Computing) - TASUS (First Workshop on Techniques and Applications for Sustainable Ultrascale Computing Systems) - UCHPC (7th Workshop on Un Conventional High Performance Computing) and VHPC (9th Workshop on Virtualization in High-Performance Cloud Computing.
This book constitutes the proceedings of the Third International Workshop on Foundational and Practical Aspects of Resource Analysis, FOPARA 2013, held in Bertinoro, Italy, in August 2013. The 9 papers presented in this volume were carefully reviewed and selected from 12 submissions. They deal with traditional approaches to complexity analysis, differential privacy, and probabilistic analysis of programs.
It is widely acknowledged that the cost of validation and testing comprises a s- nificant percentage of the overall development costs for electronic systems today, and is expected to escalate sharply in the future. Many studies have shown that up to 70% of the design development time and resources are spent on functional verification. Functional errors manifest themselves very early in the design flow, and unless they are detected up front, they can result in severe consequence- both financially and from a safety viewpoint. Indeed, several recent instances of high-profile functional errors (e. g. , the Pentium FDIV bug) have resulted in - creased attention paid to verifying the functional correctness of designs. Recent efforts have proposed augmenting the traditional RTL simulation-based validation methodology with formal techniques in an attempt to uncover hard-to-find c- ner cases, with the goal of trying to reach RTL functional verification closure. However, what is often not highlighted is the fact that in spite of the tremendous time and effort put into such efforts at the RTL and lower levels of abstraction, the complexity of contemporary embedded systems makes it difficult to guarantee functional correctness at the system level under all possible operational scenarios. The problem is exacerbated in current System-on-Chip (SOC) design meth- ologies that employ Intellectual Property (IP) blocks composed of processor cores, coprocessors, and memory subsystems. Functional verification becomes one of the major bottlenecks in the design of such systems.
This is a new type of edited volume in the Frontiers in Electronic Testing book series devoted to recent advances in electronic circuits testing. The book is a comprehensive elaboration on important topics which capture major research and development efforts today. "Hot" topics of current interest to test technology community have been selected, and the authors are key contributors in the corresponding topics.
Inthe?eldofformalmethodsincomputerscience,concurrencytheoryisreceivinga constantlyincreasinginterest.Thisisespeciallytrueforprocessalgebra.Althoughit had been originally conceived as a means for reasoning about the semantics of c- current programs, process algebraic formalisms like CCS, CSP, ACP, ?-calculus, and their extensions (see, e.g., [154,119,112,22,155,181,30]) were soon used also for comprehendingfunctionaland nonfunctionalaspects of the behaviorof com- nicating concurrent systems. The scienti?c impact of process calculi and behavioral equivalences at the base of process algebra is witnessed not only by a very rich literature. It is in fact worth mentioningthe standardizationprocedurethat led to the developmentof the process algebraic language LOTOS [49], as well as the implementation of several modeling and analysis tools based on process algebra, like CWB [70] and CADP [93], some of which have been used in industrial case studies. Furthermore, process calculi and behavioral equivalencesare by now adopted in university-levelcourses to teach the foundations of concurrent programming as well as the model-driven design of concurrent, distributed, and mobile systems. Nevertheless, after 30 years since its introduction, process algebra is rarely adopted in the practice of software development. On the one hand, its technica- ties often obfuscate the way in which systems are modeled. As an example, if a process term comprises numerous occurrences of the parallel composition operator, it is hard to understand the communicationscheme among the varioussubterms. On the other hand, process algebra is perceived as being dif?cult to learn and use by practitioners, as it is not close enough to the way they think of software systems.
This proceedings is a representation of decades of reasearch, teaching and application in the field. Image Processing, Fusion and Information Technology areas, Digital radio Communication, Wimax, Electrical engg, VLSI approach to processor design, embedded systems design are dealt in detail through models and illustrative techniques.
Information Systems Development (ISD) progresses rapidly, continually creating new challenges for the professionals involved. New concepts, approaches and techniques of systems development emerge constantly in this field. Progress in ISD comes from research as well as from practice. This conference will discuss issues pertaining to information systems development (ISD) in the inter-networked digital economy. Participants will include researchers, both experienced and novice, from industry and academia, as well as students and practitioners. Themes will include methods and approaches for ISD; ISD education; philosophical, ethical, and sociological aspects of ISD; as well as specialized tracks such as: distributed software development, ISD and knowledge management, ISD and electronic business / electronic government, ISD in public sector organizations, IOS.
Promptly growing demand for telecommunication services and information interchange has led to the fact that communication became one of the most dynamical branches of an infrastructure of a modern society. The book introduces to the bases of classical MDP theory; problems of a finding optimal in models are investigated and various problems of improvement of characteristics of traditional and multimedia wireless communication networks are considered together with both classical and new methods of theory MDP which allow defining optimal strategy of access in teletraffic systems. The book will be useful to specialists in the field of telecommunication systems and also to students and post-graduate students of corresponding specialties.
The proceedings of the 5th International Workshop on Parallel Tools for High Performance Computing provide an overview on supportive software tools and environments in the fields of System Management, Parallel Debugging and Performance Analysis. In the pursuit to maintain exponential growth for the performance of high performance computers the HPC community is currently targeting Exascale Systems. The initial planning for Exascale already started when the first Petaflop system was delivered. Many challenges need to be addressed to reach the necessary performance. Scalability, energy efficiency and fault-tolerance need to be increased by orders of magnitude. The goal can only be achieved when advanced hardware is combined with a suitable software stack. In fact, the importance of software is rapidly growing. As a result, many international projects focus on the necessary software.
Logic circuits are becoming increasingly susceptible to probabilistic behavior caused by external radiation and process variation. In addition, inherently probabilistic quantum- and nano-technologies are on the horizon as we approach the limits of CMOS scaling. Ensuring the reliability of such circuits despite the probabilistic behavior is a key challenge in IC design---one that necessitates a fundamental, probabilistic reformulation of synthesis and testing techniques. This monograph will present techniques for analyzing, designing, and testing logic circuits with probabilistic behavior.
This book constitutes the second part of the refereed proceedings of the International Conference on Life System Modeling and Simulation, LSMS 2014, and of the International Conference on Intelligent Computing for Sustainable Energy and Environment, ICSEE 2014, held in Shanghai, China, in September 2014. The 159 revised full papers presented in the three volumes of CCIS 461-463 were carefully reviewed and selected from 572 submissions. The papers of this volume are organized in topical sections on advanced neural network theory and algorithms; advanced evolutionary computing theory and algorithms, such as particle swarm optimization, differential evolution, ant colonies, artificial life, artificial immune systems and genetic algorithm; fuzzy, neural, and fuzzy-neuro hybrids; intelligent modeling, monitoring, and control of complex nonlinear systems; intelligent modeling and simulation of climate change; communication and control for distributed networked systems.
Written for those who want to develop their knowledge of requirements engineering process, whether practitioners or students. Using the latest research and driven by practical experience from industry, Requirements Engineering gives useful hints to practitioners on how to write and structure requirements. It explains the importance of Systems Engineering and the creation of effective solutions to problems. It describes the underlying representations used in system modeling and introduces the UML2, and considers the relationship between requirements and modeling. Covering a generic multi-layer requirements process, the book discusses the key elements of effective requirements management. The latest version of DOORS (Version 7) - a software tool which serves as an enabler of a requirements management process - is also introduced to the reader here. Additional material and links are available at: http://www.requirementsengineering.info
Network monitoring serves as the basis for a wide scope of network, engineering and management operations. Precise network monitoring involves inspecting every packet traversing in a network. However, this is not feasible with future high-speed networks, due to significant overheads of processing, storing, and transferring measured data. Network Monitoring in High Speed Networks presents accurate measurement schemes from both traffic and performance perspectives, and introduces adaptive sampling techniques for various granularities of traffic measurement. The techniques allow monitoring systems to control the accuracy of estimations, and adapt sampling probability dynamically according to traffic conditions. The issues surrounding network delays for practical performance monitoring are discussed in the second part of this book. Case studies based on real operational network traces are provided throughout this book. Network Monitoring in High Speed Networks is designed as a secondary text or reference book for advanced-level students and researchers concentrating on computer science and electrical engineering. Professionals working within the networking industry will also find this book useful.
Platform Based Design at the Electronic System Level presents a multi-faceted view of the challenges facing the electronic industry in the development and integration of complex heterogeneous systems, including both hardware and software components. It analyses and proposes solutions related to the provision of integration platforms by System on Chip and Integrated Platform providers in light of the needs and requirements expressed by the system companies: they are the users of such platforms, which they apply to develop their next-generation products. This is the first book to examine ESL from perspectives of system developer, platform provider and Electronic Design Automation.
A quality-driven design and verification flow for digital systems is developed and presented in Quality-Driven SystemC Design. Two major enhancements characterize the new flow: First, dedicated verification techniques are integrated which target the different levels of abstraction. Second, each verification technique is complemented by an approach to measure the achieved verification quality. The new flow distinguishes three levels of abstraction (namely system level, top level and block level) and can be incorporated in existing approaches. After reviewing the preliminary concepts, in the following chapters the three levels for modeling and verification are considered in detail. At each level the verification quality is measured. In summary, following the new design and verification flow a high overall quality results.
Performance of Web Services provides innovative techniques to improve the performance of Web Services, as well as QoS (Quality of Service) requirements. This includes Qos performance, reliability and security. The author presents two levels of Web Services: the "kernel" (ithe SOAP engine which delivers messages from one point to another through various networks), and the "server side" (which processes heavy load / requests). The primary objective of this book is execution of applications delivered in a timely fashion. Case studies and examples are provided throughout this book.
This book describes a model-based development approach for globally-asynchronous locally-synchronous distributed embedded controllers. This approach uses Petri nets as modeling formalism to create platform and network independent models supporting the use of design automation tools. To support this development approach, the Petri nets class in use is extended with time-domains and asynchronous-channels. The authors' approach uses models not only providing a better understanding of the distributed controller and improving the communication among the stakeholders, but also to be ready to support the entire lifecycle, including the simulation, the verification (using model-checking tools), the implementation (relying on automatic code generators), and the deployment of the distributed controller into specific platforms. Uses a graphical and intuitive modeling formalism supported by design automation tools; Enables verification, ensuring that the distributed controller was correctly specified; Provides flexibility in the implementation and maintenance phases to achieve desired constraints (high performance, low power consumption, reduced costs), enabling porting to different platforms using different communication nodes, without changing the underlying behavioral model.
Practical Programming in the Cell Broadband Engine offers a unique programming guide for the Cell Broadband Engine, demonstrating a large number of real-life programs to identify and solve problems in engineering, logic design, VLSI CAD, number-theory, graph-theory, computational geometry, image processing, and other subjects. Key features include:
This book constitutes the joint refereed proceedings of the 14th International Conference on Next Generation Wired/Wireless Advanced Networks and Systems, NEW2AN 2014, and the 7th Conference on Internet of Things and Smart Spaces, ruSMART 2014, held in St. Petersburg, Russia, in August 2014. The total of 67 papers was carefully reviewed and selected for inclusion in this book. The 15 papers selected from ruSMART are organized in topical sections named: smart spaces core technologies, smart spaces for geo-location and e-tourism apps, smart space supporting technologies, and video solutions for smart spaces. The 52 papers from NEW2AN deal with the following topics: advances in wireless networking, ad hoc networks and enhanced services, sensor- and machine-type communication, networking architectures and their modeling, traffic analysis and prediction, analytical methods for performance evaluation, materials for future communications, generation and analysis of signals, business aspects of networking, progress on upper layers and implementations, modeling methods and tools, techniques, algorithms, and control problems, photonics and optics, and signals and their processing.
The general trend of modern network devices towards greater intelligence and programmability is accelerating the development of systems that are increasingly autonomous and to a certain degree self-managing. Examples range from router scripting environments to fully programmable server blades. This has opened up a new field of computer science research, reflected in this new volume. This selection of contributions to the first ever international workshop on network-embedded management applications (NEMA) features six papers selected from submissions to the workshop, held in October 2010 at Niagara Falls, Canada. They represent a wide cross-section of the current work in this vital field of inquiry. Covering a diversity of perspectives, the volume's dual structure first of all examines the 'enablers' for NEMAs-the platforms, frameworks, and development environments which facilitate the evolution of network-embedded management and applications. The second section of the book covers network-embedded applications that might both empower and benefit from such enabling platforms. These papers cover topics ranging from deciding where to best place management control functions inside a network to a discussion of how multi-core hardware processors can be leveraged for traffic filtering applications. The section concludes with an analysis of a delay-tolerant network application in the context of the 'One Laptop per Child' program. There is a growing recognition that it is vital to make network operation and administration as easy as possible to contain operational expenses and cope with ever shorter control cycles. This volume provides researchers in the field with the very latest in current thinking.
This book constitutes the refereed proceedings of the 20th International Conference on Parallel and Distributed Computing, Euro-Par 2014, held in Porto, Portugal, in August 2014. The 68 revised full papers presented were carefully reviewed and selected from 267 submissions. The papers are organized in 15 topical sections: support tools environments; performance prediction and evaluation; scheduling and load balancing; high-performance architectures and compilers; parallel and distributed data management; grid, cluster and cloud computing; green high performance computing; distributed systems and algorithms; parallel and distributed programming; parallel numerical algorithms; multicore and manycore programming; theory and algorithms for parallel computation; high performance networks and communication; high performance and scientific applications; and GPU and accelerator computing.
This book constitutes the refereed proceedings of the 21st International Conference on Analytical and Stochastic Modelling Techniques and Applications, ASMTA 2014, held in Budapest, Hungary, in June/July 2014. The 18 papers presented were carefully reviewed and selected from 27 submissions. The papers discuss the latest developments in analytical, numerical and simulation algorithms for stochastic systems, including Markov processes, queueing networks, stochastic Petri nets, process algebras, game theory, etc.
Computing power performance was important at times when hardware was still expensive, because hardware had to be put to the best use. Later on this criterion was no longer critical, since hardware had become inexpensive. Meanwhile, however, people have realized that performance again plays a significant role, because of the major drain on system resources involved in developing complex applications. This book distinguishes between three levels of performance optimization: the system level, application level and business processes level. On each, optimizations can be achieved and cost-cutting potentials can be identified. The book presents the relevant theoretical background and measuring methods as well as proposed solutions. An evaluation of network monitors and checklists rounds out the work.
Haptics technology is being used more and more in different applications, such as in computer games for increased immersion, in surgical simulators to create a realistic environment for training of surgeons, in surgical robotics due to safety issues and in mobile phones to provide feedback from user action. The existence of these applications highlights a clear need to understand performance metrics for haptic interfaces and their implications on device design, use and application. Performance Metrics for Haptic Interfaces aims at meeting this need by establishing standard practices for the evaluation of haptic interfaces and by identifying significant performance metrics. Towards this end, a combined physical and psychophysical experimental methodology is presented. Firstly, existing physical performance measures and device characterization techniques are investigated and described in an illustrative way. Secondly, a wide range of human psychophysical experiments are reviewed and the appropriate ones are applied to haptic interactions. The psychophysical experiments are unified as a systematic and complete evaluation method for haptic interfaces. Finally, synthesis of both evaluation methods is discussed. The metrics provided in this state-of-the-art volume will guide readers in evaluating the performance of any haptic interface. The generic methodology will enable researchers to experimentally assess the suitability of a haptic interface for a specific purpose, to characterize and compare devices quantitatively and to identify possible improvement strategies in the design of a system.
Introductory textbook covering all the main features of the 'web programming' languages PHP and MySQL together with detailed examples that will enable readers (whether students on a taught course or independent learners) to use them to create their own applications or understand existing ones. A particular focus is the use of PHP to generate MySQL commands from a script as it is executed. Each chapter includes aims, a summary and practical exercises (with solutions) to support learning. Chapters are designed to stand alone as far as possible, so that they can be studied independently of the rest of the text by those with some previous knowledge of the languages. There is a comprehensive glossary of technical terms, together with extensive appendices for quick reference of language features. |
You may like...
Systems Analysis And Design In A…
John Satzinger, Robert Jackson, …
Hardcover
(1)
Cases on Lean Thinking Applications in…
Eduardo Guilherme Satolo, Robisom Damasceno Calado
Hardcover
R5,991
Discovery Miles 59 910
Loose Leaf for Fundamentals of Electric…
Charles Alexander, Matthew Sadiku
Loose-leaf
R5,176
Discovery Miles 51 760
|