![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > General theory of computing > Systems analysis & design
"The Testing Network" presents an integrated approach to testing based on cutting-edge methodologies, processes and tools in today's IT context. It means complex network-centric applications to be tested in heterogeneous IT infrastructures and in multiple test environments (also geographically distributed). The added-value of this book is the in-depth explanation of all processes and relevant methodologies and tools to address this complexity. Main aspects of testing are explained using TD/QC - the world-leader test platform. This up-to-date know-how is based on real-life IT experiences gained in large-scale projects of companies operating worldwide. The book is abundantly illustrated to better show all technical aspects of modern testing in a national and international context. The author has a deep expertise by designing and giving testing training in large companies using the above-mentioned tools and processes. "The Testing Network" is a unique synthesis of core test topics applied in real-life.
This book came into being inthe form oflecture notes for thesubject Infor- tion technology management (IT management) at the Twente University inthe Netherlands. Since 1995 this subject is part of the Master's degree of the course Business Management and Information Technology. Over a decade of teaching, this bookdevelopedinto what it istoday. The book gives an idea of how organizations should organize their - formationandcommunicationtechnologyfacilitiesinordertobeabletosay"IT does not matter." Management and the organization of IT are only conveniences within day-to-day operations and enablers, for organizations that want to supply other products and services. The book has the following starting points: (a) The IT support of products and services of organizations makes fu- tional and performance demandsontheIT facilities. In order to beable tomeettheserequirementsoptimally, anITarchitectureisrequired.The IT services and products are supplied within this architecture. (b) Controlling IT is part of normal operational management. This means that: -at setting up the IT facilities the principles of logistics and operations management apply; -the information, neededfor controlling a process, makes demandson the set-up of the information service process. The question is: -whether someone is authorized to supplythe data; -whether the data correspondswith thephysically present objects and -whether the given data is correct and complete. (c) A distinction is made between both the IT demand and the IT supply organization. Both organizations have to be set up. Methods indicate, xi xii Preface which processes have to be in place in these organizations and each of these processes has ?nancial, personnel, legal and security aspects.
The book provides complete coverage of fundamental IP networking in Java. It introduces the concepts behind TCP/IP and UDP and their intended use and purpose; gives complete coverage of Java networking APIs, includes an extended discussion of advanced server design, so that the various design principles and tradeoffs concerned are discussed and equips the reader with analytic queuing-theory tools to evaluate design alternatives; covers UDP multicasting, and covers multi-homed hosts, leading the reader to understand the extra programming steps and design considerations required in such environments. After reading this book the reader will have an advanced knowledge of fundamental network design and programming concepts in the Java language, enabling them to design and implement distributed applications with advanced features and to predict their performance. Special emphasis is given to the scalable I/O facilities of Java 1.4 as well as complete treatments of multi-homing and UDP both unicast and multicast.
Fundamental Problems in Computing is in honor of Professor Daniel J. Rosenkrantz, a distinguished researcher in Computer Science. Professor Rosenkrantz has made seminal contributions to many subareas of Computer Science including formal languages and compilers, automata theory, algorithms, database systems, very large scale integrated systems, fault-tolerant computing and discrete dynamical systems. For many years, Professor Rosenkrantz served as the Editor-in-Chief of the Journal of the Association for Computing Machinery (JACM), a very prestigious archival journal in Computer Science. His contributions to Computer Science have earned him many awards including the Fellowship from ACM and the ACM SIGMOD Contributions Award.
Stochastic discrete-event systems (SDES) capture the randomness in choices due to activity delays and the probabilities of decisions. This book delivers a comprehensive overview on modeling with a quantitative evaluation of SDES. It presents an abstract model class for SDES as a pivotal unifying result and details important model classes. The book also includes nontrivial examples to explain real-world applications of SDES.
The authors have here put together the first reference on all aspects of testing and validating service-oriented architectures. With contributions by leading academic and industrial research groups it offers detailed guidelines for the actual validation process. Readers will find a comprehensive survey of state-of-the-art approaches as well as techniques and tools to improve the quality of service-oriented applications. It also includes references and scenarios for future research and development.
The use of parallel programming and architectures is essential for simulating and solving problems in modern computational practice. There has been rapid progress in microprocessor architecture, interconnection technology and software devel- ment, which are in?uencing directly the rapid growth of parallel and distributed computing. However, in order to make these bene?ts usable in practice, this dev- opment must be accompanied by progress in the design, analysis and application aspects of parallel algorithms. In particular, new approaches from parallel num- ics are important for solving complex computational problems on parallel and/or distributed systems. The contributions to this book are focused on topics most concerned in the trends of today's parallel computing. These range from parallel algorithmics, progr- ming, tools, network computing to future parallel computing. Particular attention is paid to parallel numerics: linear algebra, differential equations, numerical integ- tion, number theory and their applications in computer simulations, which together form the kernel of the monograph. We expect that the book will be of interest to scientists working on parallel computing, doctoral students, teachers, engineers and mathematicians dealing with numerical applications and computer simulations of natural phenomena.
Euro-Par is an annual series of international conferences dedicated to the p- motion and advancementof allaspects of parallelcomputing. The major themes can be divided into four broad categories: theory, high-performance, cluster and grid,distributedandmobilecomputing.Thesecategoriescomprise14topicsthat focus on particular issues. The objective of Euro-Paris to provide a forum within which to promote the development of parallel computing both as an industrial technique and an a- demic discipline, extending the frontier of both the state of the art and the state of practice. The main audience for and participants in Euro-Par are researchers inacademicdepartments,governmentlaboratories,andindustrialorganizations. Euro-Par2010 was the 16th conference in the Euro-Parseries, and was or- nizedbythe Institute forHigh-PerformanceComputingandNetworking(ICAR) of the Italian National Research Council (CNR), in Ischia, Italy. Previous Euro- ParconferencestookplaceinStockholm,Lyon,Passau,Southampton,Toulouse, Munich, Manchester, Padderborn, Klagenfurt, Pisa, Lisbon, Dresden, Rennes, Las Palmas, and Delft. Next year the conference will take place in Bordeaux, France. More information on the Euro-Par conference series and organization is available on the wesite http://www.europar.org. As mentioned before, the conference was organized in 14 topics. The paper review process for each topic was managed and supervised by a committee of at least four persons: a Global Chair, a Local Chair, and two members. Some speci?c topics with a high number of submissions were managed by a larger committeewithmoremembers.The?naldecisionsontheacceptanceorrejection ofthesubmitted papersweremadein ameeting ofthe ConferenceCo-chairsand Local Chairs of the topics.
Suitable for a one- or two-semester undergraduate or beginning graduate course in computer science and computer engineering, Computer Organization, Design, and Architecture, Fifth Edition presents the operating principles, capabilities, and limitations of digital computers to enable the development of complex yet efficient systems. With 11 new sections and four revised sections, this edition takes students through a solid, up-to-date exploration of single- and multiple-processor systems, embedded architectures, and performance evaluation. See What's New in the Fifth Edition Expanded coverage of embedded systems, mobile processors, and cloud computing Material for the "Architecture and Organization" part of the 2013 IEEE/ACM Draft Curricula for Computer Science and Engineering Updated commercial machine architecture examples The backbone of the book is a description of the complete design of a simple but complete hypothetical computer. The author then details the architectural features of contemporary computer systems (selected from Intel, MIPS, ARM, Motorola, Cray and various microcontrollers, etc.) as enhancements to the structure of the simple computer. He also introduces performance enhancements and advanced architectures including networks, distributed systems, GRIDs, and cloud computing. Computer organization deals with providing just enough details on the operation of the computer system for sophisticated users and programmers. Often, books on digital systems' architecture fall into four categories: logic design, computer organization, hardware design, and system architecture. This book captures the important attributes of these four categories to present a comprehensive text that includes pertinent hardware, software, and system aspects.
Parallel and distributed computation has been gaining a great lot of attention in the last decades. During this period, the advances attained in computing and communication technologies, and the reduction in the costs of those technolo gies, played a central role in the rapid growth of the interest in the use of parallel and distributed computation in a number of areas of engineering and sciences. Many actual applications have been successfully implemented in various plat forms varying from pure shared-memory to totally distributed models, passing through hybrid approaches such as distributed-shared memory architectures. Parallel and distributed computation differs from dassical sequential compu tation in some of the following major aspects: the number of processing units, independent local dock for each unit, the number of memory units, and the programming model. For representing this diversity, and depending on what level we are looking at the problem, researchers have proposed some models to abstract the main characteristics or parameters (physical components or logical mechanisms) of parallel computers. The problem of establishing a suitable model is to find a reasonable trade-off among simplicity, power of expression and universality. Then, be able to study and analyze more precisely the behavior of parallel applications."
Euro-Par is an annual series of international conferences dedicated to the p- motion and advancement of all aspects of parallel and distributed computing. th Euro-Par 2009 was the 15 edition in this conference series. Througout the years, the Euro-Par conferences have always attracted high-quality submissions and have become one of the established conferences in the area of parallel and distributed processing. Built upon the success of the annual conferences and in order to accommodate the needs of special interest groups (among the conf- ence participants), starting from 2006, a series of workshopsin conjunction with the Euro-Par main conference have been organized. This was the ?fth year in which workshops were organized within the Euro-Par conference format. The workshops focus on advanced specialized topics in parallel and d- tributed computing. These topics re?ect new scienti?c and technological dev- opments. While the community for such new and speci?c developments is still small and the topics have yet to become mature, the Euro-Par conference o?ers a platform in the form of a workshop to exchange ideas and discuss cooperation opportunities. The workshops in the past four years have been very successful. The number ofworkshopproposalsandthenumberof?nallyacceptedworkshopshavegra- ally increasedsince 2006.In 2008, nine workshopswereorganizedin conjunction with the main Euro-Par conference. In 2009, there were again nine workshop
Human-Systems Integration: From Virtual to Tangible Subject Guide: Ergonomics and Human Factors This book is an attempt to better formalize a systemic approach to human-systems integration (HSI). Good HSI is a matter of maturity... it takes time to mature. It takes time for a human being to become autonomous, and then mature! HSI is a matter of human-machine teaming, where human-machine cooperation and coordination are crucial. We cannot think engineering design without considering people and organizations that go with it. We also cannot think new technology, new organizations, and new jobs without considering change management. More specifically, this book is a follow-up of previous contributions in human-centered design and practice in the development of virtual prototypes that requires progressive operational tangibility toward HSI. The book discusses flexibility in design and operations, tangibility of software-intensive systems, virtual human-centered design, increasingly autonomous complex systems, human factors and ergonomics of sociotechnical systems, systems integration, and changed management in digital organizations. The book will be of interest to industry, academia, those involved with systems engineering, human factors, and the broader public.
This volume comprises the edited proceedings of the second CoreGRID Integration Workshop, CGIW'2006, held October 2006 in Krakow, Poland. A "Network of Excellence" funded by the European Commission 's Sixth Framework Program, CoreGRID aims to strengthen and advance scientific and technological excellence in the area of Grid and Peer-to-Peer technologies by bringing together a critical mass of well-established researchers from 41 European research institutions. Designed for a professional audience of industry practitioners and researchers, the volume is also suitable for advanced-level students in computer science.
This book intends to inculcate the innovative ideas for the scheduling aspect in distributed computing systems. Although the models in this book have been designed for distributed systems, the same information is applicable for any type of system. The book will dramatically improve the design and management of the processes for industry professionals. It deals exclusively with the scheduling aspect, which finds little space in other distributed operating system books. Structured for a professional audience composed of researchers and practitioners in industry, this book is also suitable as a reference for graduate-level students.
The Second International Workshop on Traffic Monitoring and Analysis (TMA 2010) was an initiative of the COST Action IC0703 "Data Traffic Monitoring and Analysis: Theory, Techniques, Tools and Applications for the Future Networks" (http: // www.tma-portal.eu/cost-tma-action). The COST program is an intergovernmental framework for European cooperation in science and technology, promoting the coordination of nationally funded research on a European level. Each COST Action aims at reducing the fragmentation in - search and opening the European research area to cooperation worldwide. Traffic monitoring and analysis (TMA) is nowadays an important research topic within the field of computer networks. It involves many research groups worldwide that are collectively advancing our understanding of the Internet. The importance of TMA research is motivated by the fact that modern packet n- works are highly complex and ever-evolving objects. Understanding, developing and managing such environments is difficult and expensive in practice. Traffic monitoring is a key methodology for understanding telecommunication technology and improving its operation, and the recent advances in this field suggest that evolved TMA-based techniques can play a key role in the operation of real networks.
Welcome to the proceedings of the 19th International Workshop on Power and TimingModeling, OptimizationandSimulation, PATMOS2009.Overtheyears, PATMOShasevolvedintoanimportantEuropeanevent, whereresearchersfrom both industry and academia discuss and investigate the emerging challenges in future and contemporary applications, design methodologies, and tools required for the development of the upcoming generations of integrated circuits and s- tems. PATMOS 2009 was organized by TU Delft, The Netherlands, with sp- sorship by the NIRICT Design Lab and Cadence Design Systems, and technical co-sponsorshipbytheIEEE.Furtherinformationabouttheworkshopisavailable athttp: //ens.ewi.tudelft.nl/patmos09. The technical programof PATMOS 2009 contained state-of-the-arttechnical contributions, three invited keynotes, and a special session on SystemC-AMS Extensions. The technical program focused on timing, performance, and power consumption, as well as architectural aspects with particular emphasis on m- eling, design, characterization, analysis, and optimization in the nanometer era. The Technical Program Committee, with the assistance of additional expert reviewers, selected the 36 papers presented at PATMOS. The papers were - ganized into 7 oral sessions (with a total of 26 papers) and 2 poster sessions (with a total of 10 papers). As is customary for the PATMOS workshops, full papers were required for review, and a minimum of three reviews were received per manuscr
Companies and institutions depend more than ever on the availability of their Information Technology, and most mission critical business processes are IT-based. Business Continuity is the ability to do business under any circumstances and is an essential requirement faced by modern companies. Both concepts - High Availability and Disaster Recovery - are realized by redundant systems. This book presents requirements, concepts, and realizations of redundant systems on all abstraction levels, and all given examples refer to UNIX and Linux Systems.
Since its original inception back in 1989 the Web has changed into an environment where Web applications range from small-scale information dissemination applications, often developed by non-IT professionals, to large-scale, commercial, enterprise-planning and scheduling applications, developed by multidisciplinary teams of people with diverse skills and backgrounds and using cutting-edge, diverse technologies. As an engineering discipline, Web engineering must provide principles, methodologies and frameworks to help Web professionals and researchers develop applications and manage projects effectively. Mendes and Mosley have selected experts from numerous areas in Web engineering, who contribute chapters where important concepts are presented and then detailed using real industrial case studies. After an introduction into the discipline itself and its intricacies, the contributions range from Web effort estimation, productivity benchmarking and conceptual and model-based application development methodologies, to other important principles such as usability, reliability, testing, process improvement and quality measurement. This is the first book that looks at Web engineering from a measurement perspective. The result is a self-containing, comprehensive overview detailing the role of measurement and metrics within the context of Web engineering. This book is ideal for professionals and researchers who want to know how to use sound principles for the effective management of Web projects, as well as for courses at an advanced undergraduate or graduate level.
This book is a practical guide to IPv6 addressing Unix and network administrators with experience in TCP/IP(v4) but not necessarily any IPv6 knowledge. It focuses on reliable and efficient operation of IPv6 implementations available today rather than on protocol specifications. Consequently, it covers the essential concepts, using instructive and thoroughly tested examples, on how to configure, administrate, and debug IPv6 setups. These foundations are complemented by discussions of best practices and strategic considerations aimed at overall efficiency, reliability, maintainability, and interoperation.
A family of internationally popular microcontrollers, the Atmel AVR microcontroller series is a low-cost hardware development platform suitable for an educational environment. Until now, no text focused on the assembly language programming of these microcontrollers. Through detailed coverage of assembly language programming principles and techniques, Some Assembly Required: Assembly Language Programming with the AVR Microcontroller teaches the basic system capabilities of 8-bit AVR microcontrollers. The text illustrates fundamental computer architecture and programming structures using AVR assembly language. It employs the core AVR 8-bit RISC microcontroller architecture and a limited collection of external devices, such as push buttons, LEDs, and serial communications, to describe control structures, memory use and allocation, stacks, and I/O. Each chapter contains numerous examples and exercises, including programming problems. By studying assembly languages, computer scientists gain an understanding of the functionality of basic processors and how their capabilities support high level languages and applications. Exploring this connection between hardware and software, this book provides a foundation for understanding compilers, linkers, loaders, and operating systems in addition to the processors themselves.
Knowledge science is an emerging discipline resulting from the demands of a knowledge-based economy and information revolution. Explaining how to improve our knowledge-based society, Knowledge Science: Modeling the Knowledge Creation Process addresses problems in collecting, synthesizing, coordinating, and creating knowledge. The book introduces several key concepts in knowledge science: Knowledge technology, which encompasses classification, representation, modeling, identification, acquisition, searching, organization, storage, conversion, and dissemination Knowledge management, which covers three different yet related areas (knowledge assets, knowing processes, knower relations) Knowledge discovery and data mining, which combine databases, statistics, machine learning, and related areas to discover and extract valuable knowledge from large volumes of data Knowledge synthesis, knowledge justification, and knowledge construction, which are important in solving real-life problems Specialists in decision science, artificial intelligence, systems engineering, behavioral science, and management science, the book's contributors present their own original ideas, including an Oriental systems philosophy, a new episteme in the knowledge-based society, and a theory of knowledge construction. They emphasize the importance of systemic thinking for developing a better society in the current knowledge-based era.
Networks on Chip presents a variety of topics, problems and approaches with the common theme to systematically organize the on-chip communication in the form of a regular, shared communication network on chip, an NoC for short. As the number of processor cores and IP blocks integrated on a single chip is steadily growing, a systematic approach to design the communication infrastructure becomes necessary. Different variants of packed switched on-chip networks have been proposed by several groups during the past two years. This book summarizes the state of the art of these efforts and discusses the major issues from the physical integration to architecture to operating systems and application interfaces. It also provides a guideline and vision about the direction this field is moving to. Moreover, the book outlines the consequences of adopting design platforms based on packet switched network. The consequences may in fact be far reaching because many of the topics of distributed systems, distributed real-time systems, fault tolerant systems, parallel computer architecture, parallel programming as well as traditional system-on-chip issues will appear relevant but within the constraints of a single chip VLSI implementation. The book is organized in three parts. The first deals with system design and methodology issues. The second presents problems and solutions concerning the hardware and the basic communication infrastructure. Finally, the third part covers operating system, embedded software and application. However, communication from the physical to the application level is a central theme throughout the book. The book serves as an excellent reference source and may be used as a text for advanced courses on the subject.
Innovation in Manufacturing Networks A fundamental concept of the emergent business, scientific and technological paradigms ces area, innovation the ability to apply new ideas to products, processes, organizational practices and business models - is crucial for the future competitiveness of organizations in a continually increasingly globalised, knowledge-intensive marketplace. Responsiveness, agility as well as the high performance of manufacturing systems is responsible for the recent changes in addition to the call for new approaches to achieve cost-effective responsiveness at all the levels of an enterprise. Moreover, creating appropriate frameworks for exploring the most effective synergies between human potential and automated systems represents an enormous challenge in terms of processes characterization, modelling, and the development of adequate support tools. The implementation and use of Automation Systems requires an ever increasing knowledge of enabling technologies and Business Practices. Moreover, the digital and networked world will surely trigger new business practices. In this context and in order to achieve the desired effective and efficiency performance levels, it is crucial to maintain a balance between both the technical aspects and the human and social aspects when developing and applying new innovations and innovative enabling technologies. BASYS conferences have been developed and organized so as to promote the development of balanced automation systems in an attempt to address the majority of the current open issues.
Memory Architecture Exploration for Programmable Embedded Systems
addresses efficient exploration of alternative memory
architectures, assisted by a "compiler-in-the-loop" that allows
effective matching of the target application to the
processor-memory architecture. This new approach for memory
architecture exploration replaces the traditional black-box view of
the memory system and allows for aggressive co-optimization of the
programmable processor together with a customized memory system.
Grid Computing: Achievements and Prospects, the 9th edited volume of the CoreGRID series, includes selected papers from the CoreGRID Integration Workshop, held April 2008 in Heraklion-Crete, Greece. This event brings together representatives of the academic and industrial communities performing Grid research in Europe. The workshop was organized in the context of the CoreGRID Network of Excellence in order to provide a forum for the presentation and exchange of views on the latest developments in grid technology research. Grid Computing: Achievements and Prospects is designed for a professional audience, composed of researchers and practitioners in industry. This volume is also suitable for graduate-level students in computer science. |
![]() ![]() You may like...
Inversion of Geophysical Data, Volume 62
Cedric Schmelzbach
Hardcover
|