![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming
This book provides an overview of the problems involved in engineering scalable, elastic, and cost-efficient cloud computing services and describes the CloudScale method - a description of rescuing tools and the required steps to exploit these tools. It allows readers to analyze the scalability problem in detail and identify scalability anti-patterns and bottlenecks within an application. With the CloudScale method, software architects can analyze both existing and planned IT services. The method allows readers to answer questions like: * With an increasing number of users, can my service still deliver acceptable quality of service? * What if each user uses the service more intensively? Can my service still handle it with acceptable quality of service? * What if the number of users suddenly increases? Will my service still be able to handle it? * Will my service be cost-efficient? First the book addresses the importance of scalability, elasticity, and cost-efficiency as vital quality-related attributes of modern cloud computing applications. Following a brief overview of CloudScale, cloud computing applications are then introduced in detail and the aspects that need to be captured in models of such applications are discussed. In CloudScale, these aspects are captured in instances of the ScaleDL modeling language. Subsequently, the book describes the forward engineering part of CloudScale, which is applicable when developing a new service. It also outlines the reverse and reengineering parts of CloudScale, which come into play when an existing (legacy) service is modified. Lastly, the book directly focuses on the needs of both business-oriented and technical managers by providing guidance on all steps of implementing CloudScale as well as making decisions during that implementation. The demonstrators and reference projects described serve as a valuable starting point for learning from experience. This book is meant for all stakeholders interested in delivering scalable, elastic, and cost-efficient cloud computing applications: managers, product owners, software architects and developers alike. With this book, they can both see the overall picture as well as dive into issues of particular interest.
This book constitutes the refereed proceedings of the 14th IFIP WG 2.13 International Conference on Open Source Systems, OSS 2018, held in Athens, Greece, in June 2018. The 14 revised full papers and 2 short papers presented were carefully reviewed and selected from 38 submissions. The papers cover a wide range of topics in the field of free/libre open source software (FLOSS) and are organized in the following thematic sections: organizational aspects of OSS projects, OSS projects validity, mining OSS data, OSS in public administration, OSS governance, and OSS reusability.
This book presents a comprehensive study of multivariate time series with linear state space structure. The emphasis is put on both the clarity of the theoretical concepts and on efficient algorithms for implementing the theory. In particular, it investigates the relationship between VARMA and state space models, including canonical forms. It also highlights the relationship between Wiener-Kolmogorov and Kalman filtering both with an infinite and a finite sample. The strength of the book also lies in the numerous algorithms included for state space models that take advantage of the recursive nature of the models. Many of these algorithms can be made robust, fast, reliable and efficient. The book is accompanied by a MATLAB package called SSMMATLAB and a webpage presenting implemented algorithms with many examples and case studies. Though it lays a solid theoretical foundation, the book also focuses on practical application, and includes exercises in each chapter. It is intended for researchers and students working with linear state space models, and who are familiar with linear algebra and possess some knowledge of statistics.
This book highlights the advantages and disadvantages of various software development lifecycle models, and describes when to apply testing -- and when to use other, more cost-effective techniques. It also shows how to incorporate V&V techniques if your organization does not have a written procedure, and explains how to implement the inspection process.
Blockchain is a technology that transcends cryptocurrencies. There are other services in different sectors of the economy that can benefit from the trust and security that blockchains offer. For example, financial institutions are using blockchains for international money transfer, and in logistics, it has been used for supply chain management and tracking of goods. As more global companies and governments are experimenting and deploying blockchain solutions, it is necessary to compile knowledge on the best practices, strategies, and failures in order to create a better awareness of how blockchain could either support or add value to other services. Cross-Industry Use of Blockchain Technology and Opportunities for the Future provides emerging research highlighting the possibilities inherent in blockchain for different sectors of the economy and the added value blockchain can provide for the future of these different sectors. Featuring coverage on a broad range of topics such as data privacy, information sharing, and digital identity, this book is ideally designed for IT specialists, consultants, design engineers, cryptographers, service designers, researchers, academics, government officials, and industry professionals.
This volume presents the peer-reviewed proceedings of the international conference Imaging, Vision and Learning Based on Optimization and PDEs (IVLOPDE), held in Bergen, Norway, in August/September 2016. The contributions cover state-of-the-art research on mathematical techniques for image processing, computer vision and machine learning based on optimization and partial differential equations (PDEs). It has become an established paradigm to formulate problems within image processing and computer vision as PDEs, variational problems or finite dimensional optimization problems. This compact yet expressive framework makes it possible to incorporate a range of desired properties of the solutions and to design algorithms based on well-founded mathematical theory. A growing body of research has also approached more general problems within data analysis and machine learning from the same perspective, and demonstrated the advantages over earlier, more established algorithms. This volume will appeal to all mathematicians and computer scientists interested in novel techniques and analytical results for optimization, variational models and PDEs, together with experimental results on applications ranging from early image formation to high-level image and data analysis.
This book addresses agent-based computing, concentrating in particular on evolutionary multi-agent systems (EMAS), which have been developed since 1996 at the AGH University of Science and Technology in Cracow, Poland. It provides the relevant background information on and a detailed description of this computing paradigm, along with key experimental results. Readers will benefit from the insightful discussion, which primarily concerns the efficient implementation of computing frameworks for developing EMAS and similar computing systems, as well as a detailed formal model. Theoretical deliberations demonstrating that computing with EMAS always helps to find the optimal solution are also included, rounding out the coverage.
This book contains some selected papers from the International Conference on Extreme Learning Machine 2015, which was held in Hangzhou, China, December 15-17, 2015. This conference brought together researchers and engineers to share and exchange R&D experience on both theoretical studies and practical applications of the Extreme Learning Machine (ELM) technique and brain learning. This book covers theories, algorithms ad applications of ELM. It gives readers a glance of the most recent advances of ELM.
This edited volume on computational intelligence algorithms-based applications includes work presented at the International Conference on Computational Intelligence, Communications, and Business Analytics (CICBA 2017). It provides the latest research findings on the significance of computational intelligence and related application areas. It also introduces various computation platforms involving evolutionary algorithms, fuzzy logic, swarm intelligence, artificial neural networks and several other tools for solving real-world problems. It also discusses various tools that are hybrids of more than one solution framework, highlighting the theoretical aspects as well as various real-world applications.
This book gathers 12 of the most promising papers presented at the 15th International Conference on Software Engineering, Artificial Intelligence Research, Management and Applications (SERA 2017) held on June 7-9, 2017 at the University of Greenwich, London, UK. The aim of this conference was to bring together researchers and scientists, businessmen and entrepreneurs, teachers, engineers, computer users, and students to discuss the numerous fields of computer science, to share their experiences and to exchange new ideas and information in a meaningful way. The book also presents research findings regarding all aspects (theory, applications and tools) of computer and information science, and discusses the practical challenges encountered along the way and the solutions adopted to solve them.
This book describes RTL design using Verilog, synthesis and timing closure for System On Chip (SOC) design blocks. It covers the complex RTL design scenarios and challenges for SOC designs and provides practical information on performance improvements in SOC, as well as Application Specific Integrated Circuit (ASIC) designs. Prototyping using modern high density Field Programmable Gate Arrays (FPGAs) is discussed in this book with the practical examples and case studies. The book discusses SOC design, performance improvement techniques, testing and system level verification, while also describing the modern Intel FPGA/XILINX FPGA architectures and their use in SOC prototyping. Further, the book covers the Synopsys Design Compiler (DC) and Prime Time (PT) commands, and how they can be used to optimize complex ASIC/SOC designs. The contents of this book will be useful to students and professionals alike.
This book presents advances and innovations in grouping genetic algorithms, enriched with new and unique heuristic optimization techniques. These algorithms are specially designed for solving industrial grouping problems where system entities are to be partitioned or clustered into efficient groups according to a set of guiding decision criteria. Examples of such problems are: vehicle routing problems, team formation problems, timetabling problems, assembly line balancing, group maintenance planning, modular design, and task assignment. A wide range of industrial grouping problems, drawn from diverse fields such as logistics, supply chain management, project management, manufacturing systems, engineering design and healthcare, are presented. Typical complex industrial grouping problems, with multiple decision criteria and constraints, are clearly described using illustrative diagrams and formulations. The problems are mapped into a common group structure that can conveniently be used as an input scheme to specific variants of grouping genetic algorithms. Unique heuristic grouping techniques are developed to handle grouping problems efficiently and effectively. Illustrative examples and computational results are presented in tables and graphs to demonstrate the efficiency and effectiveness of the algorithms. Researchers, decision analysts, software developers, and graduate students from various disciplines will find this in-depth reader-friendly exposition of advances and applications of grouping genetic algorithms an interesting, informative and valuable resource.
Over the last four decades, computer systems have required increasingly complex software development and maintenance support. The marriage of software engineering, the application of engineering principals to produce economical and reliable software, to software development tools and methods promised to simplify software development while improving accuracy and speed, tools have evolved that use computer graphics to represent concepts that generate code from integrated design specifications. Practicing Software Engineering in the 21st Century addresses the tools and techniques utilized when developing and implementing software engineering practices into computer systems.
This book presents a design methodology that is practically applicable to the architectural design of a broad range of systems. It is based on fundamental design concepts to conceive and specify the required functional properties of a system, while abstracting from the specific implementation functions and technologies that can be chosen to build the system. Abstraction and precision are indispensable when it comes to understanding complex systems and precisely creating and representing them at a high functional level. Once understood, these concepts appear natural, self-evident and extremely powerful, since they can directly, precisely and concisely reflect what is considered essential for the functional behavior of a system. The first two chapters present the global views on how to design systems and how to interpret terms and meta-concepts. This informal introduction provides the general context for the remainder of the book. On a more formal level, Chapters 3 through 6 present the main basic design concepts, illustrating them with examples. Language notations are introduced along with the basic design concepts. Lastly, Chapters 7 to 12 discuss the more intricate basic design concepts of interactive systems by focusing on their common functional goal. These chapters are recommended to readers who have a particular interest in the design of protocols and interfaces for various systems. The didactic approach makes it suitable for graduate students who want to develop insights into and skills in developing complex systems, as well as practitioners in industry and large organizations who are responsible for the design and development of large and complex systems. It includes numerous tangible examples from various fields, and several appealing exercises with their solutions.
Online gaming is widely popular and gaining more user attention every day. Computer game industries have made considerable growth in terms of design and development, but the scarcity of hardware resources at player or client side is a major pitfall for the latest high-end multimedia games. Cloud gaming is one proposed solution, allowing the end-user to play games using a variety of platforms with less demanding hardware requirements. Emerging Technologies and Applications for Cloud-Based Gaming explores the opportunities for the gaming industry through the integration of cloud computing. Focusing on design methodologies, fundamental architectures, and the end-user experience, this publication is an essential reference source for IT specialists, game developers, researchers, and graduate-level students.
"The heart monitor alarm suddenly screamed as the patient's EKG pattern abruptly changed to ventricular fibrillation. "Code Blue " Dr. Singh screamed, "Get the crash cart in here, now " THUMP The convulsive thrash of his patient after each defibrillation attempt was beginning to be too much for Dr. Brady. THUMP He didn't know if he could stand the helpless feeling any longer. Attempt after attempt failed to resuscitate Mrs. Winter. He wanted to scream or run away, but continued in his efforts to save his patient. He prayed he was just going to wake up from this nightmare, hug his wife, and be thankful that this surrealistic scene didn't exist. But that simply wasn't going to happen. What began as sixty seconds of shock and terror, soon became forty minutes of futility. After trying everything they could think of to restore a normal heart rhythm to the patient's lifeless body, Dr. Singh called off the code. Jessie Winter was dead " This action-packed mystery will take you from the operating room to the courtroom as Dr. Brady searches for the truth behind his patient's unexpected death, and the resulting malpractice and manslaughter trials.
Sir Tony Hoare has had an enormous influence on computer science, from the Quicksort algorithm to the science of software development, concurrency and program verification. His contributions have been widely recognised: He was awarded the ACM's Turing Award in 1980, the Kyoto Prize from the Inamori Foundation in 2000, and was knighted for "services to education and computer science" by Queen Elizabeth II of England in 2000. This book presents the essence of his various works-the quest for effective abstractions-both in his own words as well as chapters written by leading experts in the field, including many of his research collaborators. In addition, this volume contains biographical material, his Turing award lecture, the transcript of an interview and some of his seminal papers. Hoare's foundational paper "An Axiomatic Basis for Computer Programming", presented his approach, commonly known as Hoare Logic, for proving the correctness of programs by using logical assertions. Hoare Logic and subsequent developments have formed the basis of a wide variety of software verification efforts. Hoare was instrumental in proposing the Verified Software Initiative, a cooperative international project directed at the scientific challenges of large-scale software verification, encompassing theories, tools and experiments. Tony Hoare's contributions to the theory and practice of concurrent software systems are equally impressive. The process algebra called Communicating Sequential Processes (CSP) has been one of the fundamental paradigms, both as a mathematical theory to reason about concurrent computation as well as the basis for the programming language occam. CSP served as a framework for exploring several ideas in denotational semantics such as powerdomains, as well as notions of abstraction and refinement. It is the basis for a series of industrial-strength tools which have been employed in a wide range of applications. This book also presents Hoare's work in the last few decades. These works include a rigorous approach to specifications in software engineering practice, including procedural and data abstractions, data refinement, and a modular theory of designs. More recently, he has worked with collaborators to develop Unifying Theories of Programming (UTP). Their goal is to identify the common algebraic theories that lie at the core of sequential, concurrent, reactive and cyber-physical computations.
This timely text/reference presents a comprehensive review of the workflow scheduling algorithms and approaches that are rapidly becoming essential for a range of software applications, due to their ability to efficiently leverage diverse and distributed cloud resources. Particular emphasis is placed on how workflow-based automation in software-defined cloud centers and hybrid IT systems can significantly enhance resource utilization and optimize energy efficiency. Topics and features: describes dynamic workflow and task scheduling techniques that work across multiple (on-premise and off-premise) clouds; presents simulation-based case studies, and details of real-time test bed-based implementations; offers analyses and comparisons of a broad selection of static and dynamic workflow algorithms; examines the considerations for the main parameters in projects limited by budget and time constraints; covers workflow management systems, workflow modeling and simulation techniques, and machine learning approaches for predictive workflow analytics. This must-read work provides invaluable practical insights from three subject matter experts in the cloud paradigm, which will empower IT practitioners and industry professionals in their daily assignments. Researchers and students interested in next-generation software-defined cloud environments will also greatly benefit from the material in the book.
Cloud service benchmarking can provide important, sometimes surprising insights into the quality of services and leads to a more quality-driven design and engineering of complex software architectures that use such services. Starting with a broad introduction to the field, this book guides readers step-by-step through the process of designing, implementing and executing a cloud service benchmark, as well as understanding and dealing with its results. It covers all aspects of cloud service benchmarking, i.e., both benchmarking the cloud and benchmarking in the cloud, at a basic level. The book is divided into five parts: Part I discusses what cloud benchmarking is, provides an overview of cloud services and their key properties, and describes the notion of a cloud system and cloud-service quality. It also addresses the benchmarking lifecycle and the motivations behind running benchmarks in particular phases of an application lifecycle. Part II then focuses on benchmark design by discussing key objectives (e.g., repeatability, fairness, or understandability) and defining metrics and measurement methods, and by giving advice on developing own measurement methods and metrics. Next, Part III explores benchmark execution and implementation challenges and objectives as well as aspects like runtime monitoring and result collection. Subsequently, Part IV addresses benchmark results, covering topics such as an abstract process for turning data into insights, data preprocessing, and basic data analysis methods. Lastly, Part V concludes the book with a summary, suggestions for further reading and pointers to benchmarking tools available on the Web. The book is intended for researchers and graduate students of computer science and related subjects looking for an introduction to benchmarking cloud services, but also for industry practitioners who are interested in evaluating the quality of cloud services or who want to assess key qualities of their own implementations through cloud-based experiments.
Creativity in Computing and DataFlow Supercomputing, the latest release in the Advances in Computers series published since 1960, presents detailed coverage of innovations in computer hardware, software, theory, design, and applications. In addition, it provides contributors with a medium in which they can explore topics in greater depth and breadth than journal articles typically allow. As a result, many articles have become standard references that continue to be of significant, lasting value in this rapidly expanding field.
The widespread use of XML in business and scientific databases has prompted the development of methodologies, techniques, and systems for effectively managing and analyzing XML data. This has increasingly attracted the attention of different research communities, including database, information retrieval, pattern recognition, and machine learning, from which several proposals have been offered to address problems in XML data management and knowledge discovery. XML Data Mining: Models, Methods, and Applications aims to collect knowledge from experts of database, information retrieval, machine learning, and knowledge management communities in developing models, methods, and systems for XML data mining. This book addresses key issues and challenges in XML data mining, offering insights into the various existing solutions and best practices for modeling, processing, analyzing XML data, and for evaluating performance of XML data mining algorithms and systems. |
You may like...
Research Anthology on Recent Trends…
Information Reso Management Association
Hardcover
R9,802
Discovery Miles 98 020
Fundamentals of Programming Languages
Chris Chancellor
Hardcover
Dark Silicon and Future On-chip Systems…
Suyel Namasudra, Hamid Sarbazi-Azad
Hardcover
R3,940
Discovery Miles 39 400
|