![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > General theory of computing > Data structures
Divided roughly into two sections, this book provides a brief history of the development of ECG along with heart rate variability (HRV) algorithms and the engineering innovations over the last decade in this area. It reviews clinical research, presents an overview of the clinical field, and the importance of heart rate variability in diagnosis. The book then discusses the use of particular ECG and HRV algorithms in the context of clinical applications.
This edited book first consolidates the results of the EU-funded EDISON project (Education for Data Intensive Science to Open New science frontiers), which developed training material and information to assist educators, trainers, employers, and research infrastructure managers in identifying, recruiting and inspiring the data science professionals of the future. It then deepens the presentation of the information and knowledge gained to allow for easier assimilation by the reader. The contributed chapters are presented in sequence, each chapter picking up from the end point of the previous one. After the initial book and project overview, the chapters present the relevant data science competencies and body of knowledge, the model curriculum required to teach the required foundations, profiles of professionals in this domain, and use cases and applications. The text is supported with appendices on related process models. The book can be used to develop new courses in data science, evaluate existing modules and courses, draft job descriptions, and plan and design efficient data-intensive research teams across scientific disciplines.
Exact eigenvalues, eigenvectors, and principal vectors of operators with infinite dimensional ranges can rarely be found. Therefore, one must approximate such operators by finite rank operators, then solve the original eigenvalue problem approximately. Serving as both an outstanding text for graduate students and as a source of current results for research scientists, Spectral Computations for Bounded Operators addresses the issue of solving eigenvalue problems for operators on infinite dimensional spaces. From a review of classical spectral theory through concrete approximation techniques to finite dimensional situations that can be implemented on a computer, this volume illustrates the marriage of pure and applied mathematics. It contains a variety of recent developments, including a new type of approximation that encompasses a variety of approximation methods but is simple to verify in practice. It also suggests a new stopping criterion for the QR Method and outlines advances in both the iterative refinement and acceleration techniques for improving the accuracy of approximations. The authors illustrate all definitions and results with elementary examples and include numerous exercises. Spectral Computations for Bounded Operators thus serves as both an outstanding text for second-year graduate students and as a source of current results for research scientists.
This collection of essays explores the different ways the insights from complexity theory can be applied to law. Complexity theory - a variant of systems theory - views law as an emergent, complex, self-organising system comprised of an interactive network of actors and systems that operate with no overall guiding hand, giving rise to complex, collective behaviour in law communications and actions. Addressing such issues as the unpredictability of legal systems, the ability of legal systems to adapt to changes in society, the importance of context, and the nature of law, the essays look to the implications of a complexity theory analysis for the study of public policy and administrative law, international law and human rights, regulatory practices in business and finance, and the practice of law and legal ethics. These are areas where law, which craves certainty, encounters unending, irresolvable complexity. This collection shows the many ways complexity theory thinking can reshape and clarify our understanding of the various problems relating to the theory and practice of law.
Fuzzy social choice theory is useful for modeling the uncertainty and imprecision prevalent in social life yet it has been scarcely applied and studied in the social sciences. Filling this gap, Application of Fuzzy Logic to Social Choice Theory provides a comprehensive study of fuzzy social choice theory. The book explains the concept of a fuzzy maximal subset of a set of alternatives, fuzzy choice functions, the factorization of a fuzzy preference relation into the "union" (conorm) of a strict fuzzy relation and an indifference operator, fuzzy non-Arrowian results, fuzzy versions of Arrow's theorem, and Black's median voter theorem for fuzzy preferences. It examines how unambiguous and exact choices are generated by fuzzy preferences and whether exact choices induced by fuzzy preferences satisfy certain plausible rationality relations. The authors also extend known Arrowian results involving fuzzy set theory to results involving intuitionistic fuzzy sets as well as the Gibbard-Satterthwaite theorem to the case of fuzzy weak preference relations. The final chapter discusses Georgescu's degree of similarity of two fuzzy choice functions.
This book collects selected contributions presented at the INdAM Workshop "Geometric Challenges in Isogeometric Analysis", held in Rome, Italy on January 27-31, 2020. It gives an overview of the forefront research on splines and their efficient use in isogeometric methods for the discretization of differential problems over complex and trimmed geometries. A variety of research topics in this context are covered, including (i) high-quality spline surfaces on complex and trimmed geometries, (ii) construction and analysis of smooth spline spaces on unstructured meshes, (iii) numerical aspects and benchmarking of isogeometric discretizations on unstructured meshes, meshing strategies and software. Given its scope, the book will be of interest to both researchers and graduate students working in the areas of approximation theory, geometric design and numerical simulation. Chapter 10 is available open access under a Creative Commons Attribution 4.0 International License via link.springer.com.
Synthetic Aperture Radar Automatic Detection Algorithms (SARADA) for Oil Spills conveys the pivotal tool required to fully comprehend the advanced algorithms in radar monitoring and detection of oil spills, particularly quantum computing and algorithms as a keystone to comprehending theories and algorithms behind radar imaging and detection of marine pollution. Bridging the gap between modern quantum mechanics and computing detection algorithms of oil spills, this book contains precise theories and techniques for automatic identification of oil spills from SAR measurements. Based on modern quantum physics, the book also includes the novel theory on radar imaging mechanism of oil spills. With the use of precise quantum simulation of trajectory movements of oil spills using a sequence of radar images, this book demonstrates the use of SARADA for contamination by oil spills as a promising novel technique. Key Features: Introduces basic concepts of a radar remote sensing. Fills a gap in the knowledge base of quantum theory and microwave remote sensing. Discusses the important aspects of oil spill imaging in radar data in relation to the quantum theory. Provides recent developments and progresses of automatic detection algorithms of oil spill from radar data. Presents 2-D oil spill radar data in 4-D images.
For computer scientists, especially those in the security field, the use of chaos has been limited to the computation of a small collection of famous but unsuitable maps that offer no explanation of why chaos is relevant in the considered contexts. Discrete Dynamical Systems and Chaotic Machines: Theory and Applications shows how to make finite machines, such as computers, neural networks, and wireless sensor networks, work chaotically as defined in a rigorous mathematical framework. Taking into account that these machines must interact in the real world, the authors share their research results on the behaviors of discrete dynamical systems and their use in computer science. Covering both theoretical and practical aspects, the book presents: Key mathematical and physical ideas in chaos theory Computer science fundamentals, clearly establishing that chaos properties can be satisfied by finite state machines Concrete applications of chaotic machines in computer security, including pseudorandom number generators, hash functions, digital watermarking, and steganography Concrete applications of chaotic machines in wireless sensor networks, including secure data aggregation and video surveillance Until the authors' recent research, the practical implementation of the mathematical theory of chaos on finite machines raised several issues. This self-contained book illustrates how chaos theory enables the study of computer security problems, such as steganalysis, that otherwise could not be tackled. It also explains how the theory reinforces existing cryptographically secure tools and schemes.
Pattern Recognition Algorithms for Data Mining addresses different pattern recognition (PR) tasks in a unified framework with both theoretical and experimental results. Tasks covered include data condensation, feature selection, case generation, clustering/classification, and rule generation and evaluation. This volume presents various theories, methodologies, and algorithms, using both classical approaches and hybrid paradigms. The authors emphasize large datasets with overlapping, intractable, or nonlinear boundary classes, and datasets that demonstrate granular computing in soft frameworks. Organized into eight chapters, the book begins with an introduction to PR, data mining, and knowledge discovery concepts. The authors analyze the tasks of multi-scale data condensation and dimensionality reduction, then explore the problem of learning with support vector machine (SVM). They conclude by highlighting the significance of granular computing for different mining tasks in a soft paradigm.
This book provides theoretical and practical knowledge on AI and swarm intelligence. It provides a methodology for EA (evolutionary algorithm)-based approach for complex adaptive systems with the integration of several meta-heuristics, e.g., ACO (Ant Colony Optimization), ABC (Artificial Bee Colony), and PSO (Particle Swarm Optimization), etc. These developments contribute towards better problem-solving methodologies in AI. The book also covers emerging uses of swarm intelligence in applications such as complex adaptive systems, reaction-diffusion computing, and diffusion-limited aggregation, etc. Another emphasis is its real-world applications. We give empirical examples from real-world problems and show that the proposed approaches are successful when addressing tasks from such areas as swarm robotics, silicon traffics, image understanding, Vornoi diagrams, queuing theory, and slime intelligence, etc. Each chapter begins with the background of the problem followed by the current state-of-the-art techniques of the field, and ends with a detailed discussion. In addition, the simulators, based on optimizers such as PSO and ABC complex adaptive system simulation, are described in detail. These simulators, as well as some source codes, are available online on the author's website for the benefit of readers interested in getting some hands-on experience of the subject. The concepts presented in this book aim to promote and facilitate the effective research in swarm intelligence approaches in both theory and practice. This book would also be of value to other readers because it covers interdisciplinary research topics that encompass problem-solving tasks in AI, complex adaptive systems, and meta-heuristics.
'The book under review is an interesting elaboration that fills the gaps in libraries for concisely written and student-friendly books about essentials in computer science ... I recommend this book for anyone who would like to study algorithms, learn a lot about computer science or simply would like to deepen their knowledge ... The book is written in very simple English and can be understood even by those with limited knowledge of the English language. It should be emphasized that, despite the fact that the book consists of many examples, mathematical formulas and theorems, it is very hard to find any mistakes, errors or typos.'zbMATHIn computer science, an algorithm is an unambiguous specification of how to solve a class of problems. Algorithms can perform calculation, data processing and automated reasoning tasks.As an effective method, an algorithm can be expressed within a finite amount of space and time and in a well-defined formal language for calculating a function. Starting from an initial state and initial input (perhaps empty), the instructions describe a computation that, when executed, proceeds through a finite number of well-defined successive states, eventually producing 'output' and terminating at a final ending state. The transition from one state to the next is not necessarily deterministic; some algorithms, known as randomized algorithms, incorporate random input.This book introduces a set of concepts in solving problems computationally such as Growth of Functions; Backtracking; Divide and Conquer; Greedy Algorithms; Dynamic Programming; Elementary Graph Algorithms; Minimal Spanning Tree; Single-Source Shortest Paths; All Pairs Shortest Paths; Flow Networks; Polynomial Multiplication, to ways of solving NP-Complete Problems, supported with comprehensive, and detailed problems and solutions, making it an ideal resource to those studying computer science, computer engineering and information technology.
This practically-focused study guide introduces the fundamentals of discrete mathematics through an extensive set of classroom-tested problems. Each chapter presents a concise introduction to the relevant theory, followed by a detailed account of common challenges and methods for overcoming these. The reader is then encouraged to practice solving such problems for themselves, by tackling a varied selection of questions and assignments of different levels of complexity. This updated second edition now covers the design and analysis of algorithms using Python, and features more than 50 new problems, complete with solutions. Topics and features: provides a substantial collection of problems and examples of varying levels of difficulty, suitable for both laboratory practical training and self-study; offers detailed solutions to each problem, applying commonly-used methods and computational schemes; introduces the fundamentals of mathematical logic, the theory of algorithms, Boolean algebra, graph theory, sets, relations, functions, and combinatorics; presents more advanced material on the design and analysis of algorithms, including Turing machines, asymptotic analysis, and parallel algorithms; includes reference lists of trigonometric and finite summation formulae in an appendix, together with basic rules for differential and integral calculus. This hands-on workbook is an invaluable resource for undergraduate students of computer science, informatics, and electronic engineering. Suitable for use in a one- or two-semester course on discrete mathematics, the text emphasizes the skills required to develop and implement an algorithm in a specific programming language.
These contributions, written by the foremost international researchers and practitioners of Genetic Programming (GP), explore the synergy between theoretical and empirical results on real-world problems, producing a comprehensive view of the state of the art in GP. Topics in this volume include: evolutionary constraints, relaxation of selection mechanisms, diversity preservation strategies, flexing fitness evaluation, evolution in dynamic environments, multi-objective and multi-modal selection, foundations of evolvability, evolvable and adaptive evolutionary operators, foundation of injecting expert knowledge in evolutionary search, analysis of problem difficulty and required GP algorithm complexity, foundations in running GP on the cloud - communication, cooperation, flexible implementation, and ensemble methods. Additional focal points for GP symbolic regression are: (1) The need to guarantee convergence to solutions in the function discovery mode; (2) Issues on model validation; (3) The need for model analysis workflows for insight generation based on generated GP solutions - model exploration, visualization, variable selection, dimensionality analysis; (4) Issues in combining different types of data. Readers will discover large-scale, real-world applications of GP to a variety of problem domains via in-depth presentations of the latest and most significant results.
Marking a distinct departure from the perspectives of frame theory and discrete transforms, this book provides a comprehensive mathematical and algorithmic introduction to wavelet theory. As such, it can be used as either a textbook or reference guide. As a textbook for graduate mathematics students and beginning researchers, it offers detailed information on the basic theory of framelets and wavelets, complemented by self-contained elementary proofs, illustrative examples/figures, and supplementary exercises. Further, as an advanced reference guide for experienced researchers and practitioners in mathematics, physics, and engineering, the book addresses in detail a wide range of basic and advanced topics (such as multiwavelets/multiframelets in Sobolev spaces and directional framelets) in wavelet theory, together with systematic mathematical analysis, concrete algorithms, and recent developments in and applications of framelets and wavelets. Lastly, the book can also be used to teach on or study selected special topics in approximation theory, Fourier analysis, applied harmonic analysis, functional analysis, and wavelet-based signal/image processing.
This book gathers selected papers presented at the International Conference on Advancements in Computing and Management (ICACM 2019). Discussing current research in the field of artificial intelligence and machine learning, cloud computing, recent trends in security, natural language processing and machine translation, parallel and distributed algorithms, as well as pattern recognition and analysis, it is a valuable resource for academics, practitioners in industry and decision-makers.
Computational Complexity of Counting and Sampling provides readers with comprehensive and detailed coverage of the subject of computational complexity. It is primarily geared toward researchers in enumerative combinatorics, discrete mathematics, and theoretical computer science. The book covers the following topics: Counting and sampling problems that are solvable in polynomial running time, including holographic algorithms; #P-complete counting problems; and approximation algorithms for counting and sampling. First, it opens with the basics, such as the theoretical computer science background and dynamic programming algorithms. Later, the book expands its scope to focus on advanced topics, like stochastic approximations of counting discrete mathematical objects and holographic algorithms. After finishing the book, readers will agree that the subject is well covered, as the book starts with the basics and gradually explores the more complex aspects of the topic. Features: Each chapter includes exercises and solutions Ideally written for researchers and scientists Covers all aspects of the topic, beginning with a solid introduction, before shifting to computational complexity's more advanced features, with a focus on counting and sampling
Teaching quantum computation and information is notoriously difficult, because it requires covering subjects from various fields of science, organizing these subjects consistently in a unified way despite their tendency to favor their specific languages, and overcoming the subjects' abstract and theoretical natures, which offer few examples of actual realizations. In this book, we have organized all the subjects required to understand the principles of quantum computation and information processing in a manner suited to physics, mathematics, and engineering courses as early as undergraduate studies.In addition, we provide a supporting package of quantum simulation software from Wolfram Mathematica, specialists in symbolic calculation software. Throughout the book's main text, demonstrations are provided that use the software package, allowing the students to deepen their understanding of each subject through self-practice. Readers can change the code so as to experiment with their own ideas and contemplate possible applications. The information in this book reflects many years of experience teaching quantum computation and information. The quantum simulation-based demonstrations and the unified organization of the subjects are both time-tested and have received very positive responses from the students who have experienced them.
A metaheuristic is a higher-level procedure designed to select a partial search algorithm that may lead to a good solution to an optimization problem, especially with incomplete or imperfect information.This unique compendium focuses on the insights of hybrid metaheuristics. It illustrates the recent researches on evolving novel hybrid metaheuristic algorithms, and prominently highlights its diverse application areas. As such, the book helps readers to grasp the essentials of hybrid metaheuristics and to address real world problems.The must-have volume serves as an inspiring read for professionals, researchers, academics and graduate students in the fields of artificial intelligence, robotics and machine learning.Related Link(s)
Originally published in 1995, Large Deviations for Performance Analysis consists of two synergistic parts. The first half develops the theory of large deviations from the beginning, through recent results on the theory for processes with boundaries, keeping to a very narrow path: continuous-time, discrete-state processes. By developing only what is needed for the applications, the theory is kept to a manageable level, both in terms of length and in terms of difficulty. Within its scope, the treatment is detailed, comprehensive and self-contained. As the book shows, there are sufficiently many interesting applications of jump Markov processes to warrant a special treatment. The second half is a collection of applications developed at Bell Laboratories. The applications cover large areas of the theory of communication networks: circuit switched transmission, packet transmission, multiple access channels, and the M/M/1 queue. Aspects of parallel computation are covered as well including, basics of job allocation, rollback-based parallel simulation, assorted priority queueing models that might be used in performance models of various computer architectures, and asymptotic coupling of processors. These applications are thoroughly analysed using the tools developed in the first half of the book.
This book takes its reader on a journey through Apache Giraph, a popular distributed graph processing platform designed to bring the power of big data processing to graph data. Designed as a step-by-step self-study guide for everyone interested in large-scale graph processing, it describes the fundamental abstractions of the system, its programming models and various techniques for using the system to process graph data at scale, including the implementation of several popular and advanced graph analytics algorithms. The book is organized as follows: Chapter 1 starts by providing a general background of the big data phenomenon and a general introduction to the Apache Giraph system, its abstraction, programming model and design architecture. Next, chapter 2 focuses on Giraph as a platform and how to use it. Based on a sample job, even more advanced topics like monitoring the Giraph application lifecycle and different methods for monitoring Giraph jobs are explained. Chapter 3 then provides an introduction to Giraph programming, introduces the basic Giraph graph model and explains how to write Giraph programs. In turn, Chapter 4 discusses in detail the implementation of some popular graph algorithms including PageRank, connected components, shortest paths and triangle closing. Chapter 5 focuses on advanced Giraph programming, discussing common Giraph algorithmic optimizations, tunable Giraph configurations that determine the system's utilization of the underlying resources, and how to write a custom graph input and output format. Lastly, chapter 6 highlights two systems that have been introduced to tackle the challenge of large scale graph processing, GraphX and GraphLab, and explains the main commonalities and differences between these systems and Apache Giraph. This book serves as an essential reference guide for students, researchers and practitioners in the domain of large scale graph processing. It offers step-by-step guidance, with several code examples and the complete source code available in the related github repository. Students will find a comprehensive introduction to and hands-on practice with tackling large scale graph processing problems using the Apache Giraph system, while researchers will discover thorough coverage of the emerging and ongoing advancements in big graph processing systems.
This book presents a consolidated survey of the vibrant field of research known as the theory of semi-feasible algorithms. This research stream perfectly showcases the richness of, and contrasts between, the central notions of complexity: running time, nonuniform complexity, lowness, and NP-hardness. Research into semi-feasible computation has already developed a rich set of tools, yet is young enough to have an abundance of fresh, open issues. Being essentially self-contained, the book requires neither great mathematical maturity nor an extensive background in computational complexity theory or in computer science in general. Newcomers are introduced to the field systematically and guided to the frontiers of current research. Researchers already active in the field will appreciate the book as a valuable source of reference.
This textbook is a second edition of Evolutionary Algorithms for Solving Multi-Objective Problems, significantly expanded and adapted for the classroom. The various features of multi-objective evolutionary algorithms are presented here in an innovative and student-friendly fashion, incorporating state-of-the-art research. The book disseminates the application of evolutionary algorithm techniques to a variety of practical problems. It contains exhaustive appendices, index and bibliography and links to a complete set of teaching tutorials, exercises and solutions.
Reservoir operation is a multi-objective optimization problem, and is traditionally solved with dynamic programming (DP) and stochastic dynamic programming (SDP) algorithms. The thesis presents novel algorithms for optimal reservoir operation, named nested DP (nDP), nested SDP (nSDP), nested reinforcement learning (nRL) and their multi-objective (MO) variants, correspondingly MOnDP, MOnSDP and MOnRL. The idea is to include a nested optimization algorithm into each state transition, which reduces the initial problem dimension and alleviates the curse of dimensionality. These algorithms can solve multi-objective optimization problems, without significantly increasing the algorithm complexity or the computational expenses. It can additionally handle dense and irregular variable discretization. All algorithms are coded in Java and were tested on the case study of the Knezevo reservoir in the Republic of Macedonia. Nested optimization algorithms are embedded in a cloud application platform for water resources modeling and optimization. The platform is available 24/7, accessible from everywhere, scalable, distributed, interoperable, and it creates a real-time multiuser collaboration platform. This thesis contributes with new and more powerful algorithms for an optimal reservoir operation and cloud application platform. All source codes are available for public use and can be used by researchers and practitioners to further advance the mentioned areas.
The field of evolutionary computation is expanding dramatically, fueled by the vast investment that reflects the value of applying its techniques. Culling material from the Handbook of Evolutionary Computation, Evolutionary Computation 1: Basic Algorithms and Operators contains up-to-date information on algorithms and operators used in evolutionary computing. This volume discusses the basic ideas that underlie the main paradigms of evolutionary algorithms, evolution strategies, evolutionary programming, and genetic programming. It is intended to be used by individual researchers, teachers, and students working and studying in this expanding field.
Algorithms and Applications for Academic Search, Recommendation and Quantitative Association Rule Mining presents novel algorithms for academic search, recommendation and association rule mining that have been developed and optimized for different commercial as well as academic purpose systems. Along with the design and implementation of algorithms, a major part of the work presented in the book involves the development of new systems both for commercial as well as for academic use. In the first part of the book the author introduces a novel hierarchical heuristic scheme for re-ranking academic publications retrieved from standard digital libraries. The scheme is based on the hierarchical combination of a custom implementation of the term frequency heuristic, a time-depreciated citation score and a graph-theoretic computed score that relates the paper's index terms with each other. In order to evaluate the performance of the introduced algorithms, a meta-search engine has been designed and developed that submits user queries to standard digital repositories of academic publications and re-ranks the top-n results using the introduced hierarchical heuristic scheme. In the second part of the book the design of novel recommendation algorithms with application in different types of e-commerce systems are described. The newly introduced algorithms are a part of a developed Movie Recommendation system, the first such system to be commercially deployed in Greece by a major Triple Play services provider. The initial version of the system uses a novel hybrid recommender (user, item and content based) and provides daily recommendations to all active subscribers of the provider (currently more than 30,000). The recommenders that we are presenting are hybrid by nature, using an ensemble configuration of different content, user as well as item-based recommenders in order to provide more accurate recommendation results. The final part of the book presents the design of a quantitative association rule mining algorithm. Quantitative association rules refer to a special type of association rules of the form that antecedent implies consequent consisting of a set of numerical or quantitative attributes. The introduced mining algorithm processes a specific number of user histories in order to generate a set of association rules with a minimally required support and confidence value. The generated rules show strong relationships that exist between the consequent and the antecedent of each rule, representing different items that have been consumed at specific price levels. This research book will be of appeal to researchers, graduate students, professionals, engineers and computer programmers. |
You may like...
The Definitive Guide to Linux Network…
Nathan Yocom, John Turner, …
Paperback
|