![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming
Businesses must constantly adapt to a dynamically changing environment that requires choosing an adaptive and dynamic information architecture that has the flexibility to support both changes in the business environment and changes in technology. In general, information systems reengineering has the objective of extracting the contents, data structures, and flow of data and process contained within existing legacy systems in order to reconstitute them into a new form for subsequent implementation. Information Systems Reengineering for Modern Business Systems: ERP, Supply Chain and E-Commerce Management Solutions covers different techniques that could be used in industry in order to reengineer business processes and legacy systems into more flexible systems capable of supporting modern trends such as Enterprise Resource Planning (ERP), supply chain management systems and e-commerce. This reference book also covers other issues related to the reengineering of legacy systems, which include risk management and obsolescence management of requirements.
This book contains some selected papers from the International Conference on Extreme Learning Machine 2015, which was held in Hangzhou, China, December 15-17, 2015. This conference brought together researchers and engineers to share and exchange R&D experience on both theoretical studies and practical applications of the Extreme Learning Machine (ELM) technique and brain learning. This book covers theories, algorithms ad applications of ELM. It gives readers a glance of the most recent advances of ELM.
This book presents a design methodology that is practically applicable to the architectural design of a broad range of systems. It is based on fundamental design concepts to conceive and specify the required functional properties of a system, while abstracting from the specific implementation functions and technologies that can be chosen to build the system. Abstraction and precision are indispensable when it comes to understanding complex systems and precisely creating and representing them at a high functional level. Once understood, these concepts appear natural, self-evident and extremely powerful, since they can directly, precisely and concisely reflect what is considered essential for the functional behavior of a system. The first two chapters present the global views on how to design systems and how to interpret terms and meta-concepts. This informal introduction provides the general context for the remainder of the book. On a more formal level, Chapters 3 through 6 present the main basic design concepts, illustrating them with examples. Language notations are introduced along with the basic design concepts. Lastly, Chapters 7 to 12 discuss the more intricate basic design concepts of interactive systems by focusing on their common functional goal. These chapters are recommended to readers who have a particular interest in the design of protocols and interfaces for various systems. The didactic approach makes it suitable for graduate students who want to develop insights into and skills in developing complex systems, as well as practitioners in industry and large organizations who are responsible for the design and development of large and complex systems. It includes numerous tangible examples from various fields, and several appealing exercises with their solutions.
This book describes RTL design using Verilog, synthesis and timing closure for System On Chip (SOC) design blocks. It covers the complex RTL design scenarios and challenges for SOC designs and provides practical information on performance improvements in SOC, as well as Application Specific Integrated Circuit (ASIC) designs. Prototyping using modern high density Field Programmable Gate Arrays (FPGAs) is discussed in this book with the practical examples and case studies. The book discusses SOC design, performance improvement techniques, testing and system level verification, while also describing the modern Intel FPGA/XILINX FPGA architectures and their use in SOC prototyping. Further, the book covers the Synopsys Design Compiler (DC) and Prime Time (PT) commands, and how they can be used to optimize complex ASIC/SOC designs. The contents of this book will be useful to students and professionals alike.
This volume presents the peer-reviewed proceedings of the international conference Imaging, Vision and Learning Based on Optimization and PDEs (IVLOPDE), held in Bergen, Norway, in August/September 2016. The contributions cover state-of-the-art research on mathematical techniques for image processing, computer vision and machine learning based on optimization and partial differential equations (PDEs). It has become an established paradigm to formulate problems within image processing and computer vision as PDEs, variational problems or finite dimensional optimization problems. This compact yet expressive framework makes it possible to incorporate a range of desired properties of the solutions and to design algorithms based on well-founded mathematical theory. A growing body of research has also approached more general problems within data analysis and machine learning from the same perspective, and demonstrated the advantages over earlier, more established algorithms. This volume will appeal to all mathematicians and computer scientists interested in novel techniques and analytical results for optimization, variational models and PDEs, together with experimental results on applications ranging from early image formation to high-level image and data analysis.
This book presents advances and innovations in grouping genetic algorithms, enriched with new and unique heuristic optimization techniques. These algorithms are specially designed for solving industrial grouping problems where system entities are to be partitioned or clustered into efficient groups according to a set of guiding decision criteria. Examples of such problems are: vehicle routing problems, team formation problems, timetabling problems, assembly line balancing, group maintenance planning, modular design, and task assignment. A wide range of industrial grouping problems, drawn from diverse fields such as logistics, supply chain management, project management, manufacturing systems, engineering design and healthcare, are presented. Typical complex industrial grouping problems, with multiple decision criteria and constraints, are clearly described using illustrative diagrams and formulations. The problems are mapped into a common group structure that can conveniently be used as an input scheme to specific variants of grouping genetic algorithms. Unique heuristic grouping techniques are developed to handle grouping problems efficiently and effectively. Illustrative examples and computational results are presented in tables and graphs to demonstrate the efficiency and effectiveness of the algorithms. Researchers, decision analysts, software developers, and graduate students from various disciplines will find this in-depth reader-friendly exposition of advances and applications of grouping genetic algorithms an interesting, informative and valuable resource.
Web technologies have become a vital element within educational, professional, and social settings as they have the potential to improve performance and productivity across organizations. Artificial Intelligence Technologies and the Evolution of Web 3.0 brings together emergent research and best practices surrounding the effective usage of Web 3.0 technologies in a variety of environments. Featuring the latest technologies and applications across industries, this publication is a vital reference source for academics, researchers, students, and professionals who are interested in new ways to use intelligent web technologies within various settings.
Over the last four decades, computer systems have required increasingly complex software development and maintenance support. The marriage of software engineering, the application of engineering principals to produce economical and reliable software, to software development tools and methods promised to simplify software development while improving accuracy and speed, tools have evolved that use computer graphics to represent concepts that generate code from integrated design specifications. Practicing Software Engineering in the 21st Century addresses the tools and techniques utilized when developing and implementing software engineering practices into computer systems.
"The heart monitor alarm suddenly screamed as the patient's EKG pattern abruptly changed to ventricular fibrillation. "Code Blue " Dr. Singh screamed, "Get the crash cart in here, now " THUMP The convulsive thrash of his patient after each defibrillation attempt was beginning to be too much for Dr. Brady. THUMP He didn't know if he could stand the helpless feeling any longer. Attempt after attempt failed to resuscitate Mrs. Winter. He wanted to scream or run away, but continued in his efforts to save his patient. He prayed he was just going to wake up from this nightmare, hug his wife, and be thankful that this surrealistic scene didn't exist. But that simply wasn't going to happen. What began as sixty seconds of shock and terror, soon became forty minutes of futility. After trying everything they could think of to restore a normal heart rhythm to the patient's lifeless body, Dr. Singh called off the code. Jessie Winter was dead " This action-packed mystery will take you from the operating room to the courtroom as Dr. Brady searches for the truth behind his patient's unexpected death, and the resulting malpractice and manslaughter trials.
The only way to stop a hacker is to think like one
Systems Software: Essential Concepts provides students with an accessible introduction to the knowledge and fundamentals that are necessary to understand software and hardware. The text also reviews implementation techniques to familiarize students with more complex software, such as operating systems and compilers, and prepare them to take more advance courses within the discipline. The book is divided into two key topical areas: compiler fundamentals and the basic mechanisms and data structures required to support operating systems. In the compiler section, students learn about the runtime environment, how to implement a scanner and a symbol table, and how to implement parsing and code generation for a virtual machine. To emphasize practical application, students are challenged to implement a small compiler. In the operating system domain, students gain an understanding of the interrupt mechanism, process and thread implementation, and process synchronization. Featuring a modern and practical approach, Systems Software is an ideal resource for courses in system programming, systems software, software development, and assembly language. It can also serve as a supplementary material for introductory operating system and compiler courses.
Cloud service benchmarking can provide important, sometimes surprising insights into the quality of services and leads to a more quality-driven design and engineering of complex software architectures that use such services. Starting with a broad introduction to the field, this book guides readers step-by-step through the process of designing, implementing and executing a cloud service benchmark, as well as understanding and dealing with its results. It covers all aspects of cloud service benchmarking, i.e., both benchmarking the cloud and benchmarking in the cloud, at a basic level. The book is divided into five parts: Part I discusses what cloud benchmarking is, provides an overview of cloud services and their key properties, and describes the notion of a cloud system and cloud-service quality. It also addresses the benchmarking lifecycle and the motivations behind running benchmarks in particular phases of an application lifecycle. Part II then focuses on benchmark design by discussing key objectives (e.g., repeatability, fairness, or understandability) and defining metrics and measurement methods, and by giving advice on developing own measurement methods and metrics. Next, Part III explores benchmark execution and implementation challenges and objectives as well as aspects like runtime monitoring and result collection. Subsequently, Part IV addresses benchmark results, covering topics such as an abstract process for turning data into insights, data preprocessing, and basic data analysis methods. Lastly, Part V concludes the book with a summary, suggestions for further reading and pointers to benchmarking tools available on the Web. The book is intended for researchers and graduate students of computer science and related subjects looking for an introduction to benchmarking cloud services, but also for industry practitioners who are interested in evaluating the quality of cloud services or who want to assess key qualities of their own implementations through cloud-based experiments.
This book addresses agent-based computing, concentrating in particular on evolutionary multi-agent systems (EMAS), which have been developed since 1996 at the AGH University of Science and Technology in Cracow, Poland. It provides the relevant background information on and a detailed description of this computing paradigm, along with key experimental results. Readers will benefit from the insightful discussion, which primarily concerns the efficient implementation of computing frameworks for developing EMAS and similar computing systems, as well as a detailed formal model. Theoretical deliberations demonstrating that computing with EMAS always helps to find the optimal solution are also included, rounding out the coverage.
How do we define the nature of our business, gather everything that we know about it, and then centralize our information in one, easily accessed place within the organization? Breslin and McGann call such knowledge our ways of working and the place where it will be found a business knowledge repository. All of a company's accumulated operations data, its manuals and procedures, its records of compliance with myriad regulations, its audits, disaster recovery plans--are essential information that today's management needs at its fingertips, and information that tomorroW's management must be sure can easily be found. Breslin and McGann show clearly and comprehensively how business knowledge repositories can be established and maintained, what should go into them and how to get it out, who should have access, and all of the other details that management needs to make the most of this valuable resource and means of doing business. An essential study and guide for management at upper levels in all types of organizations, both public and private. Breslin and McGann show that once an organization's knowledge of itself is formulated into its ways of working, its so-called object orientation makes it easily maintained. The repository approach to organizing and consolidating knowledge makes it possible for all of its potential users to access it easily, without having to go to one source for one thing they need and to another for another thing, a tedious and costly procedure in many organizations that have allowed their information and knowledge resources to not only grow but become duplicated as well. The repository approach also makes it possible for management to organize and access information by job functions, and to make it available to employees more easily in training situations. Regulators and auditors are also more easily served. As a result, CFOs will find their annual audit and various compliance fees considerably reduced. Breslin and McGann's book is thus a blueprint for the creation of knowledge repositories and a discussion of how graphical communication between information systems creators and their client end users can be made to flow smoothly and efficiently.
This two-part book puts the spotlight on how a real-time kernel works using Micrium's C/OS-III kernel as a reference. Part I includes an overview of the operation of real-time kernels, and walks through various aspects of C/OS-III implementation and usage. Part II provides application examples (using the versatile Renesas YRDKRX62N Evaluation Board, available separately) that enable readers to rapidly develop their own prototypes. This book is written for serious embedded systems programmers, consultants, hobbyists, and students interested in understanding the inner workings of a real-time kernel. C/OS-III is not just a great learning platform, but also a full commercial-grade software package, ready to be part of a wide range of products. C/OS-III is a highly portable, ROMable, scalable, preemptive real-time, multitasking kernel designed specifically to address the demanding requirements of today 's embedded systems. C/OS-III is the successor to the highly popular C/OS-II real-time kernel but can use most of C/OS-II 's ports with minor modifications. Some of the features of C/OS-III are: Preemptive multitasking with round-robin scheduling of tasks at
the same priority
The widespread use of XML in business and scientific databases has prompted the development of methodologies, techniques, and systems for effectively managing and analyzing XML data. This has increasingly attracted the attention of different research communities, including database, information retrieval, pattern recognition, and machine learning, from which several proposals have been offered to address problems in XML data management and knowledge discovery. XML Data Mining: Models, Methods, and Applications aims to collect knowledge from experts of database, information retrieval, machine learning, and knowledge management communities in developing models, methods, and systems for XML data mining. This book addresses key issues and challenges in XML data mining, offering insights into the various existing solutions and best practices for modeling, processing, analyzing XML data, and for evaluating performance of XML data mining algorithms and systems.
As systems become more prevalent and more complex, resilient adaptive systems are crucial when systems are needed in environments where change is the rule rather than the exception. Technological Innovations in Adaptive and Dependable Systems: Advancing Models and Concepts provides high quality, effective approaches to design, develop, maintain, evaluate, and benchmark adaptive and dependable systems that are built to sustain quality of service and experience despite the occurrence of potentially significant and sudden changes or failures in their infrastructure and surrounding environments. Providing academicians, practitioners, and researchers with insight, this book contains useful software and hardware aspects, conceptual models, applied and theoretical approaches, paradigms, and other technological innovations.
This edited volume on computational intelligence algorithms-based applications includes work presented at the International Conference on Computational Intelligence, Communications, and Business Analytics (CICBA 2017). It provides the latest research findings on the significance of computational intelligence and related application areas. It also introduces various computation platforms involving evolutionary algorithms, fuzzy logic, swarm intelligence, artificial neural networks and several other tools for solving real-world problems. It also discusses various tools that are hybrids of more than one solution framework, highlighting the theoretical aspects as well as various real-world applications. |
You may like...
Exploring Zynq MPSoC - With PYNQ and…
Crockett H Louise, Northcote David, …
Hardcover
R1,884
Discovery Miles 18 840
8088 and 8086 Microprocessors, The…
Walter Triebel, Avtar Singh
Paperback
R2,724
Discovery Miles 27 240
|