![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer hardware & operating systems
Today, it is not uncommon for practices and hospitals to be on their second or third EHR and/or contemplating a transition from the traditional on-premise model to a cloud-based system. As a follow-up to Complete Guide and Toolkit to Successful EHR Adoption ( (c)2011 HIMSS), this book builds on the best practices of the first edition, fast-forwarding to the latest innovations that are currently leveraged and adopted by providers and hospitals. We examine the role that artificial intelligence (AI) is now playing in and around EHR technology. We also address the advances in analytics and deep learning (also known as deep structured or hierarchical learning) and explain this topic in practical ways for even the most novice reader to comprehend and apply. The challenges of EHR to EHR migrations and data conversions will also be covered, including the use of the unethical practice of data blocking used as a tactic by some vendors to hold data hostage. Further, we explore innovations related to interoperability, cloud computing, cyber security, and electronic patient/consumer engagement. Finally, this book will deal with what to do with aging technology and databases, which is an issue rarely considered in any of the early publications on healthcare technology. What is the proper way to retire a legacy system, and what are the legal obligations of data archiving? Though a lot has changed since the 2011 edition, many of the fundamentals remain the same and will serve as a foundation for the next generation of EHR adopters and/or those moving on to their second, third, fourth, and beyond EHRs.
Internet of Things with 8051 and ESP8266 provides a platform to get started with the Internet of Things (IoT) with 8051. This book describes programming basics and how devices interface within designed systems. It presents a unique combination of 8051 with ESP8266 and I/O devices for IoT applications supported by case studies to provide the solutions to real-time problems. The programs and circuits have been tested on real hardware and explore different areas in IoT applications. Divided into four sections, it explains the customized boards for IoT applications followed by the means by which 8051 and ESP8266 interface with I/O devices. It spans levels from basic to advanced interfacing with special devices, server design, and data logging with different platforms. Features: Covers how I/O devices interface with 8051 and ESP8266 Explains the basic concepts of interfacing complexity using applications with examples Provides hands-on practice exercises with 8051 and ESP8266 for IoT applications Discusses both case studies and programming tests on real hardware during industrial and student projects Reviews the integration of smart devices with IoT Internet of Things with 8051 and ESP8266 is intended for senior undergraduate and graduate students in electrical and electronics engineering, but anyone with an interest in the professional curriculum of electrical and electronics engineering will find this book a welcome addition to their collection.
This book covers both theoretical and practical results for graph polynomials. Graph polynomials have been developed for measuring combinatorial graph invariants and for characterizing graphs. Various problems in pure and applied graph theory or discrete mathematics can be treated and solved efficiently by using graph polynomials. Graph polynomials have been proven useful areas such as discrete mathematics, engineering, information sciences, mathematical chemistry and related disciplines.
A Guide to the National Initiative for Cybersecurity Education (NICE) Cybersecurity Workforce Framework (2.0) presents a comprehensive discussion of the tasks, knowledge, skill, and ability (KSA) requirements of the NICE Cybersecurity Workforce Framework 2.0. It discusses in detail the relationship between the NICE framework and the NIST's cybersecurity framework (CSF), showing how the NICE model specifies what the particular specialty areas of the workforce should be doing in order to ensure that the CSF's identification, protection, defense, response, or recovery functions are being carried out properly. The authors construct a detailed picture of the proper organization and conduct of a strategic infrastructure security operation, describing how these two frameworks provide an explicit definition of the field of cybersecurity. The book is unique in that it is based on well-accepted standard recommendations rather than presumed expertise. It is the first book to align with and explain the requirements of a national-level initiative to standardize the study of information security. Moreover, it contains knowledge elements that represent the first fully validated and authoritative body of knowledge (BOK) in cybersecurity. The book is divided into two parts: The first part is comprised of three chapters that give you a comprehensive understanding of the structure and intent of the NICE model, its various elements, and their detailed contents. The second part contains seven chapters that introduce you to each knowledge area individually. Together, these parts help you build a comprehensive understanding of how to organize and execute a cybersecurity workforce definition using standard best practice.
Although interest in Spatial Decision Support Systems (SDSS) continues to grow rapidly in a wide range of disciplines, students, planners, managers, and the research community have lacked a book that covers the fundamentals of SDSS along with the advanced design concepts required for building SDSS. Filling this need, Spatial Decision Support Systems: Principles and Practices provides a comprehensive examination of the various aspects of SDSS evolution, components, architecture, and implementation. It integrates research from a variety of disciplines, including the geosciences, to supply a complete overview of SDSS technologies and their application from an interdisciplinary perspective. This groundbreaking reference provides thorough coverage of the roots of SDSS. It explains the core principles of SDSS, how to use them in various decision making contexts, and how to design and develop them using readily available enabling technologies and commercial tools. The book consists of four major parts, each addressing different topic areas in SDSS: Presents an introduction to SDSS and the evolution of SDSS Covers the essential and optional components of SDSS Focuses on the design and implementation of SDSS Reviews SDSS applications from various domains and disciplines-investigating current challenges and future directions The text includes numerous detailed case studies, example applications, and methods for tailoring SDSS to your work environment. It also integrates sample code segments throughout. Addressing the technical and organizational challenges that affect the success or failure of SDSS, the book concludes by considering future directions of this rapidly emerging field of study.
This book constitutes the refereed proceedings of the 1st IFIP Entertainment Computing Symposium held in Milan, Italy on September 7-10, 2008. The IFIP series publishes state-of-the-art results in the sciences and technologies of information and communication. The scope of the series includes: foundations of computer science; software theory and practice; education; computer applications in technology; communication systems; systems modeling and optimization; information systems; computers and society; computer systems technology; security and protection in information processing systems; artificial intelligence; and human-computer interaction. Proceedings and post-proceedings of refereed international conferences in computer science and interdisciplinary fields are featured. These results often precede journal publication and represent the most current research. The principal aim of the IFIP series is to encourage education and the dissemination and exchange of information about all aspects of computing.
This textbook covers the design of electronic systems from the ground up, from drawing and CAD essentials to recycling requirements. Chapter by chapter, it deals with the challenges any modern system designer faces: The design process and its fundamentals, such as technical drawings and CAD, electronic system levels, assembly and packaging issues and appliance protection classes, reliability analysis, thermal management and cooling, electromagnetic compatibility (EMC), all the way to recycling requirements and environmental-friendly design principles. "This unique book provides fundamental, complete, and indispensable information regarding the design of electronic systems. This topic has not been addressed as complete and thorough anywhere before. Since the authors are world-renown experts, it is a foundational reference for today's design professionals, as well as for the next generation of engineering students." Dr. Patrick Groeneveld, Synopsys Inc.
Businesses today are faced with a highly competitive market and fast-changing technologies. In order to meet demanding customers' needs, they rely on high quality software. A new field of study, soft computing techniques, is needed to estimate the efforts invested in component-based software. Component-Based Systems: Estimating Efforts Using Soft Computing Techniques is an important resource that uses computer-based models for estimating efforts of software. It provides an overview of component-based software engineering, while addressing uncertainty involved in effort estimation and expert opinions. This book will also instruct the reader how to develop mathematical models. This book is an excellent source of information for students and researchers to learn soft computing models, their applications in software management, and will help software developers, managers, and those in the industry to apply soft computing techniques to estimate efforts.
This book presents task-scheduling techniques for emerging complex parallel architectures including heterogeneous multi-core architectures, warehouse-scale datacenters, and distributed big data processing systems. The demand for high computational capacity has led to the growing popularity of multicore processors, which have become the mainstream in both the research and real-world settings. Yet to date, there is no book exploring the current task-scheduling techniques for the emerging complex parallel architectures. Addressing this gap, the book discusses state-of-the-art task-scheduling techniques that are optimized for different architectures, and which can be directly applied in real parallel systems. Further, the book provides an overview of the latest advances in task-scheduling policies in parallel architectures, and will help readers understand and overcome current and emerging issues in this field.
Learn, prepare, and practice for Red Hat RHCE 8 EX294 exam success with this Cert Guide from Pearson IT Certification, a leader in IT certification learning. * Master RHCE 8 EX294 exam topics * Assess your knowledge with chapter-ending quizzes * Review key concepts with step-by-step exercises, code examples, and complete labs after every chapter, designed to closely resemble the exam * Practice with realistic exam questions from four full practice exams: two printed in the book and two more on the companion website * Supplement your learning with an hour of video mentoring Red Hat RHCE 8 (EX294) Cert Guide is the authoritative exam study guide. Leading Red Hat and Linux consultant, trainer, and author Sander van Vugt walks you through the RHCE EX 294 objectives so you have a full study resource. He shares preparation hints and test-taking tips, helping you identify areas of weakness and improve both your conceptual knowledge and hands-on skills. Material is presented in a concise manner, focusing on increasing your understanding, retention of exam topics, and hands-on practice so you can feel ready to take the exam. The book presents you with an organized test preparation routine through the use of proven series elements and techniques. Exam topic lists make referencing easy. Chapter-ending Exam Preparation Tasks help you drill on key concepts you must know thoroughly. Review questions help you assess your knowledge, and a final preparation chapter guides you through tools and resources to help you craft your final study plan. The end of chapter labs are an essential part of learning each topic and cover each chapter's key concepts so you can test yourself and put your knowledge to work. The companion website contains two additional practice exams, flashcards that test you on the glossary terms in the book, an hour of video mentoring from the author, and study resources. Well regarded for its level of detail, assessment features, and challenging review questions and exercises, this study guide helps you master the concepts and techniques that will enable you to succeed on the exam the first time. The study guide helps you master all the topics on the RHCE 8 exam, including * Understanding configuration management * Installing Ansible and setting up managed environments * Using ad hoc commands * Getting started with playbooks * Working with variables and facts * Using task control * Deploying files * Using Ansible roles * Using Ansible in large environments * Troubleshooting Ansible * Managing software with Ansible * Managing users, processes, tasks, and storage
This is a monumental reference for the theory and practice of computer security. Comprehensive in scope, this text covers applied and practical elements, theory, and the reasons for the design of applications and security techniques. It covers both the management and the engineering issues of computer security. It provides excellent examples of ideas and mechanisms that demonstrate how disparate techniques and principles are combined in widely-used systems. This book is acclaimed for its scope, clear and lucid writing, and its combination of formal and theoretical aspects with real systems, technologies, techniques, and policies.
Artificial Intelligence (AI), when incorporated with machine learning and deep learning algorithms, has a wide variety of applications today. This book focuses on the implementation of various elementary and advanced approaches in AI that can be used in various domains to solve real-time decision-making problems. The book focuses on concepts and techniques used to run tasks in an automated manner. It discusses computational intelligence in the detection and diagnosis of clinical and biomedical images, covers the automation of a system through machine learning and deep learning approaches, presents data analytics and mining for decision-support applications, and includes case-based reasoning, natural language processing, computer vision, and AI approaches in real-time applications. Academic scientists, researchers, and students in the various domains of computer science engineering, electronics and communication engineering, and information technology, as well as industrial engineers, biomedical engineers, and management, will find this book useful. By the end of this book, you will understand the fundamentals of AI. Various case studies will develop your adaptive thinking to solve real-time AI problems. Features Includes AI-based decision-making approaches Discusses computational intelligence in the detection and diagnosis of clinical and biomedical images Covers automation of systems through machine learning and deep learning approaches and its implications to the real world Presents data analytics and mining for decision-support applications Offers case-based reasoning
This book discusses Change Management Impact Analysis and how this method is used to analysis the risks and benefits of a change management initiative when it pertains to obtaining critical insight into how the change management program budget should be allotted. The process also offers useful indicators for what areas within the system should be monitored during the change management process. This book presents theoretical analysis of practical implications and surveys, along with analysis. It covers the functions aimed at identifying various stakeholders associated with the software such as requirement component, design component, and class component. The book talks about the interrelationship between the change and the effects on the rest of the system and dives deeper to include the critical role that the analysis places on the existing multiple functions such as estimating the development costs, the project overhead costs, cost for the modification of the system, and system strength or detecting errors in the system during the process. Case studies are also included to help researchers and practitioners to absorb the material presented. This book is useful to graduate students, researchers, academicians, institutions, and professionals that interested in exploring the areas of Impact Analysis.
Modern computing relies on future and emergent technologies which have been conceived via interaction between computer science, engineering, chemistry, physics and biology. This highly interdisciplinary book presents advances in the fields of parallel, distributed and emergent information processing and computation. The book represents major breakthroughs in parallel quantum protocols, elastic cloud servers, structural properties of interconnection networks, internet of things, morphogenetic collective systems, swarm intelligence and cellular automata, unconventionality in parallel computation, algorithmic information dynamics, localized DNA computation, graph-based cryptography, slime mold inspired nano-electronics and cytoskeleton computers. Features Truly interdisciplinary, spanning computer science, electronics, mathematics and biology Covers widely popular topics of future and emergent computing technologies, cloud computing, parallel computing, DNA computation, security and network analysis, cryptography, and theoretical computer science Provides unique chapters written by top experts in theoretical and applied computer science, information processing and engineering From Parallel to Emergent Computing provides a visionary statement on how computing will advance in the next 25 years and what new fields of science will be involved in computing engineering. This book is a valuable resource for computer scientists working today, and in years to come.
Information security has a major gap when cryptography is implemented. Cryptographic algorithms are well defined, key management schemes are well known, but the actual deployment is typically overlooked, ignored, or unknown. Cryptography is everywhere. Application and network architectures are typically well-documented but the cryptographic architecture is missing. This book provides a guide to discovering, documenting, and validating cryptographic architectures. Each chapter builds on the next to present information in a sequential process. This approach not only presents the material in a structured manner, it also serves as an ongoing reference guide for future use.
The book shows how to use Planning by Design (PxD) for developing working models to any type of subject area. Section 1 describes the nature of planning in general, the formula of planning, the features that make it systematic, the essence of PxD, and developing and using the working model. Section 2 demonstrates personal application of creative planning to real life cases and practical working models on different subject areas. The book provides a general planning "master guide" that shows how to develop a working model of any definable subject matter. This objective will be accomplished by introducing the concepts, the process, and the methodology of PxD.
The EU's General Data Protection Regulation created the position of corporate Data Protection Officer (DPO), who is empowered to ensure the organization is compliant with all aspects of the new data protection regime. Organizations must now appoint and designate a DPO. The specific definitions and building blocks of the data protection regime are enhanced by the new General Data Protection Regulation and therefore the DPO will be very active in passing the message and requirements of the new data protection regime throughout the organization. This book explains the roles and responsiblies of the DPO, as well as highlights the potential cost of getting data protection wrong.
Melvin Greer and Kevin Jackson have assembled a comprehensive guide to industry-specific cybersecurity threats and provide a detailed risk management framework required to mitigate business risk associated with the adoption of cloud computing. This book can serve multiple purposes, not the least of which is documenting the breadth and severity of the challenges that today's enterprises face, and the breadth of programmatic elements required to address these challenges. This has become a boardroom issue: Executives must not only exploit the potential of information technologies, but manage their potential risks. Key Features * Provides a cross-industry view of contemporary cloud computing security challenges, solutions, and lessons learned * Offers clear guidance for the development and execution of industry-specific cloud computing business and cybersecurity strategies * Provides insight into the interaction and cross-dependencies between industry business models and industry-specific cloud computing security requirements
Big Data of Complex Networks presents and explains the methods from the study of big data that can be used in analysing massive structural data sets, including both very large networks and sets of graphs. As well as applying statistical analysis techniques like sampling and bootstrapping in an interdisciplinary manner to produce novel techniques for analyzing massive amounts of data, this book also explores the possibilities offered by the special aspects such as computer memory in investigating large sets of complex networks. Intended for computer scientists, statisticians and mathematicians interested in the big data and networks, Big Data of Complex Networks is also a valuable tool for researchers in the fields of visualization, data analysis, computer vision and bioinformatics. Key features: Provides a complete discussion of both the hardware and software used to organize big data Describes a wide range of useful applications for managing big data and resultant data sets Maintains a firm focus on massive data and large networks Unveils innovative techniques to help readers handle big data Matthias Dehmer received his PhD in computer science from the Darmstadt University of Technology, Germany. Currently, he is Professor at UMIT - The Health and Life Sciences University, Austria, and the Universitat der Bundeswehr Munchen. His research interests are in graph theory, data science, complex networks, complexity, statistics and information theory. Frank Emmert-Streib received his PhD in theoretical physics from the University of Bremen, and is currently Associate professor at Tampere University of Technology, Finland. His research interests are in the field of computational biology, machine learning and network medicine. Stefan Pickl holds a PhD in mathematics from the Darmstadt University of Technology, and is currently a Professor at Bundeswehr Universitat Munchen. His research interests are in operations research, systems biology, graph theory and discrete optimization. Andreas Holzinger received his PhD in cognitive science from Graz University and his habilitation (second PhD) in computer science from Graz University of Technology. He is head of the Holzinger Group HCI-KDD at the Medical University Graz and Visiting Professor for Machine Learning in Health Informatics Vienna University of Technology.
Enterprise Level Security: Securing Information Systems in an Uncertain World provides a modern alternative to the fortress approach to security. The new approach is more distributed and has no need for passwords or accounts. Global attacks become much more difficult, and losses are localized, should they occur. The security approach is derived from a set of tenets that form the basic security model requirements. Many of the changes in authorization within the enterprise model happen automatically. Identities and claims for access occur during each step of the computing process. Many of the techniques in this book have been piloted. These techniques have been proven to be resilient, secure, extensible, and scalable. The operational model of a distributed computer environment defense is currently being implemented on a broad scale for a particular enterprise. The first section of the book comprises seven chapters that cover basics and philosophy, including discussions on identity, attributes, access and privilege, cryptography, the cloud, and the network. These chapters contain an evolved set of principles and philosophies that were not apparent at the beginning of the project. The second section, consisting of chapters eight through twenty-two, contains technical information and details obtained by making painful mistakes and reworking processes until a workable formulation was derived. Topics covered in this section include claims-based authentication, credentials for access claims, claims creation, invoking an application, cascading authorization, federation, and content access control. This section also covers delegation, the enterprise attribute ecosystem, database access, building enterprise software, vulnerability analyses, the enterprise support desk, and network defense.
"I highly recommend Mr. Hobbs' book." - Stephen Thomas, PE, Founder and Editor of FunctionalSafetyEngineer.com Safety-critical devices, whether medical, automotive, or industrial, are increasingly dependent on the correct operation of sophisticated software. Many standards have appeared in the last decade on how such systems should be designed and built. Developers, who previously only had to know how to program devices for their industry, must now understand remarkably esoteric development practices and be prepared to justify their work to external auditors. Embedded Software Development for Safety-Critical Systems discusses the development of safety-critical systems under the following standards: IEC 61508; ISO 26262; EN 50128; and IEC 62304. It details the advantages and disadvantages of many architectural and design practices recommended in the standards, ranging from replication and diversification, through anomaly detection to the so-called "safety bag" systems. Reviewing the use of open-source components in safety-critical systems, this book has evolved from a course text used by QNX Software Systems for a training module on building embedded software for safety-critical devices, including medical devices, railway systems, industrial systems, and driver assistance devices in cars. Although the book describes open-source tools for the most part, it also provides enough information for you to seek out commercial vendors if that's the route you decide to pursue. All of the techniques described in this book may be further explored through hundreds of learned articles. In order to provide you with a way in, the author supplies references he has found helpful as a working software developer. Most of these references are available to download for free.
While aviation fatalities have thankfully fallen dramatically in recent years, the phenomena of complexity and cognitive bias have been shown to be factors in many accidents. An understanding of these phenomena promises to bring the fatality rate even lower, and a deeper understanding of commercial aircraft in the context of systems engineering will contribute to that trend. Systems Approach to the Design of Commercial Aircraft describes commercial aircraft from an advanced systems point of view, addressing complexity, cybersecurity, and systems architecting. In addition, it provides an explanation of systems engineering, describes how systems engineering forms a framework for commercial aircraft, covers how systems engineering and systems architecting relate to commercial aircraft, addresses complexity, and shows how humans fit into systems engineering and the importance for commercial aircraft. It goes onto present how cybersecurity plays an important role in the mix and how human interface fits in. The readership includes designers of aircraft, manufacturers, researchers, systems engineers, and students. Scott Jackson is a fellow of the International Council on Systems Engineering (INCOSE) and the author of Systems Engineering for Commercial Aircraft (1997 and 2015) in English and Chinese. Ricardo Moraes dos Santos is a senior systems engineer at EMBRAER S/A and an INCOSE Brazil chapter director. He works with Architecting process (Corporate) and is head of Cybersecurity and Safety (STPA Applications) at EMBRAER S/A.
Conquer today’s Windows 10–from the inside out!
In view of the growing presence and popularity of multicore and manycore processors, accelerators, and coprocessors, as well as clusters using such computing devices, the development of efficient parallel applications has become a key challenge to be able to exploit the performance of such systems. This book covers the scope of parallel programming for modern high performance computing systems. It first discusses selected and popular state-of-the-art computing devices and systems available today, These include multicore CPUs, manycore (co)processors, such as Intel Xeon Phi, accelerators, such as GPUs, and clusters, as well as programming models supported on these platforms. It next introduces parallelization through important programming paradigms, such as master-slave, geometric Single Program Multiple Data (SPMD) and divide-and-conquer. The practical and useful elements of the most popular and important APIs for programming parallel HPC systems are discussed, including MPI, OpenMP, Pthreads, CUDA, OpenCL, and OpenACC. It also demonstrates, through selected code listings, how selected APIs can be used to implement important programming paradigms. Furthermore, it shows how the codes can be compiled and executed in a Linux environment. The book also presents hybrid codes that integrate selected APIs for potentially multi-level parallelization and utilization of heterogeneous resources, and it shows how to use modern elements of these APIs. Selected optimization techniques are also included, such as overlapping communication and computations implemented using various APIs. Features: Discusses the popular and currently available computing devices and cluster systems Includes typical paradigms used in parallel programs Explores popular APIs for programming parallel applications Provides code templates that can be used for implementation of paradigms Provides hybrid code examples allowing multi-level parallelization Covers the optimization of parallel programs
GPU Parallel Program Development using CUDA teaches GPU programming by showing the differences among different families of GPUs. This approach prepares the reader for the next generation and future generations of GPUs. The book emphasizes concepts that will remain relevant for a long time, rather than concepts that are platform-specific. At the same time, the book also provides platform-dependent explanations that are as valuable as generalized GPU concepts. The book consists of three separate parts; it starts by explaining parallelism using CPU multi-threading in Part I. A few simple programs are used to demonstrate the concept of dividing a large task into multiple parallel sub-tasks and mapping them to CPU threads. Multiple ways of parallelizing the same task are analyzed and their pros/cons are studied in terms of both core and memory operation. Part II of the book introduces GPU massive parallelism. The same programs are parallelized on multiple Nvidia GPU platforms and the same performance analysis is repeated. Because the core and memory structures of CPUs and GPUs are different, the results differ in interesting ways. The end goal is to make programmers aware of all the good ideas, as well as the bad ideas, so readers can apply the good ideas and avoid the bad ideas in their own programs. Part III of the book provides pointer for readers who want to expand their horizons. It provides a brief introduction to popular CUDA libraries (such as cuBLAS, cuFFT, NPP, and Thrust),the OpenCL programming language, an overview of GPU programming using other programming languages and API libraries (such as Python, OpenCV, OpenGL, and Apple's Swift and Metal,) and the deep learning library cuDNN. |
You may like...
Clean Architecture - A Craftsman's Guide…
Robert Martin
Paperback
(1)
Edsger Wybe Dijkstra - His Life, Work…
Krzysztof R. Apt, Tony Hoare
Hardcover
R2,920
Discovery Miles 29 200
Practical TCP/IP and Ethernet Networking…
Deon Reynders, Edwin Wright
Paperback
R1,491
Discovery Miles 14 910
|