![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > General theory of computing > Data structures
This book aims to provide some insights into recently developed bio-inspired algorithms within recent emerging trends of fog computing, sentiment analysis, and data streaming as well as to provide a more comprehensive approach to the big data management from pre-processing to analytics to visualization phases. The subject area of this book is within the realm of computer science, notably algorithms (meta-heuristic and, more particularly, bio-inspired algorithms). Although application domains of these new algorithms may be mentioned, the scope of this book is not on the application of algorithms to specific or general domains but to provide an update on recent research trends for bio-inspired algorithms within a specific application domain or emerging area. These areas include data streaming, fog computing, and phases of big data management. One of the reasons for writing this book is that the bio-inspired approach does not receive much attention but shows considerable promise and diversity in terms of approach of many issues in big data and streaming. Some novel approaches of this book are the use of these algorithms to all phases of data management (not just a particular phase such as data mining or business intelligence as many books focus on); effective demonstration of the effectiveness of a selected algorithm within a chapter against comparative algorithms using the experimental method. Another novel approach is a brief overview and evaluation of traditional algorithms, both sequential and parallel, for use in data mining, in order to provide an overview of existing algorithms in use. This overview complements a further chapter on bio-inspired algorithms for data mining to enable readers to make a more suitable choice of algorithm for data mining within a particular context. In all chapters, references for further reading are provided, and in selected chapters, the author also include ideas for future research.
This book comprises select proceedings of the 7th International Conference on Data Science and Engineering (ICDSE 2021). The contents of this book focus on responsible data science. This book tries to integrate research across diverse topics related to data science, such as fairness, trust, ethics, confidentiality, transparency, and accuracy. The chapters in this book represent research from different perspectives that offer novel theoretical implications that span multiple disciplines. The book will serve as a reference resource for researchers and practitioners in academia and industry.
Although rigidity has been studied since the time of Lagrange (1788) and Maxwell (1864), it is only in the last twenty-five years that it has begun to find applications in the basic sciences. The modern era starts with Laman (1970), who made the subject rigorous in two dimensions, followed by the development of computer algorithms that can test over a million sites in seconds and find the rigid regions, and the associated pivots, leading to many applications. This workshop was organized to bring together leading researchers studying the underlying theory, and to explore the various areas of science where applications of these ideas are being implemented.
The burgeoning field of data analysis is expanding at an incredible pace due to the proliferation of data collection in almost every area of science. The enormous data sets now routinely encountered in the sciences provide an incentive to develop mathematical techniques and computational algorithms that help synthesize, interpret and give meaning to the data in the context of its scientific setting. A specific aim of this book is to integrate standard scientific computing methods with data analysis. By doing so, it brings together, in a self-consistent fashion, the key ideas from: * statistics, * time-frequency analysis, and * low-dimensional reductions The blend of these ideas provides meaningful insight into the data sets one is faced with in every scientific subject today, including those generated from complex dynamical systems. This is a particularly exciting field and much of the final part of the book is driven by intuitive examples from it, showing how the three areas can be used in combination to give critical insight into the fundamental workings of various problems. Data-Driven Modeling and Scientific Computation is a survey of practical numerical solution techniques for ordinary and partial differential equations as well as algorithms for data manipulation and analysis. Emphasis is on the implementation of numerical schemes to practical problems in the engineering, biological and physical sciences. An accessible introductory-to-advanced text, this book fully integrates MATLAB and its versatile and high-level programming functionality, while bringing together computational and data skills for both undergraduate and graduate students in scientific computing.
The five-volume set IFIP AICT 630, 631, 632, 633, and 634 constitutes the refereed proceedings of the International IFIP WG 5.7 Conference on Advances in Production Management Systems, APMS 2021, held in Nantes, France, in September 2021.*The 378 papers presented were carefully reviewed and selected from 529 submissions. They discuss artificial intelligence techniques, decision aid and new and renewed paradigms for sustainable and resilient production systems at four-wall factory and value chain levels. The papers are organized in the following topical sections: Part I: artificial intelligence based optimization techniques for demand-driven manufacturing; hybrid approaches for production planning and scheduling; intelligent systems for manufacturing planning and control in the industry 4.0; learning and robust decision support systems for agile manufacturing environments; low-code and model-driven engineering for production system; meta-heuristics and optimization techniques for energy-oriented manufacturing systems; metaheuristics for production systems; modern analytics and new AI-based smart techniques for replenishment and production planning under uncertainty; system identification for manufacturing control applications; and the future of lean thinking and practice Part II: digital transformation of SME manufacturers: the crucial role of standard; digital transformations towards supply chain resiliency; engineering of smart-product-service-systems of the future; lean and Six Sigma in services healthcare; new trends and challenges in reconfigurable, flexible or agile production system; production management in food supply chains; and sustainability in production planning and lot-sizing Part III: autonomous robots in delivery logistics; digital transformation approaches in production management; finance-driven supply chain; gastronomic service system design; modern scheduling and applications in industry 4.0; recent advances in sustainable manufacturing; regular session: green production and circularity concepts; regular session: improvement models and methods for green and innovative systems; regular session: supply chain and routing management; regular session: robotics and human aspects; regular session: classification and data management methods; smart supply chain and production in society 5.0 era; and supply chain risk management under coronavirus Part IV: AI for resilience in global supply chain networks in the context of pandemic disruptions; blockchain in the operations and supply chain management; data-based services as key enablers for smart products, manufacturing and assembly; data-driven methods for supply chain optimization; digital twins based on systems engineering and semantic modeling; digital twins in companies first developments and future challenges; human-centered artificial intelligence in smart manufacturing for the operator 4.0; operations management in engineer-to-order manufacturing; product and asset life cycle management for smart and sustainable manufacturing systems; robotics technologies for control, smart manufacturing and logistics; serious games analytics: improving games and learning support; smart and sustainable production and supply chains; smart methods and techniques for sustainable supply chain management; the new digital lean manufacturing paradigm; and the role of emerging technologies in disaster relief operations: lessons from COVID-19 Part V: data-driven platforms and applications in production and logistics: digital twins and AI for sustainability; regular session: new approaches for routing problem solving; regular session: improvement of design and operation of manufacturing systems; regular session: crossdock and transportation issues; regular session: maintenance improvement and lifecycle management; regular session: additive manufacturing and mass customization; regular session: frameworks and conceptual modelling for systems and services efficiency; regular session: optimization of production and transportation systems; regular session: optimization of supply chain agility and reconfigurability; regular session: advanced modelling approaches; regular session: simulation and optimization of systems performances; regular session: AI-based approaches for quality and performance improvement of production systems; and regular session: risk and performance management of supply chains *The conference was held online.
This book reviews research developments in diverse areas of reinforcement learning such as model-free actor-critic methods, model-based learning and control, information geometry of policy searches, reward design, and exploration in biology and the behavioral sciences. Special emphasis is placed on advanced ideas, algorithms, methods, and applications. The contributed papers gathered here grew out of a lecture course on reinforcement learning held by Prof. Jan Peters in the winter semester 2018/2019 at Technische Universitat Darmstadt. The book is intended for reinforcement learning students and researchers with a firm grasp of linear algebra, statistics, and optimization. Nevertheless, all key concepts are introduced in each chapter, making the content self-contained and accessible to a broader audience.
This book includes constructive approximation theory; it presents ordinary and fractional approximations by positive sublinear operators, and high order approximation by multivariate generalized Picard, Gauss-Weierstrass, Poisson-Cauchy and trigonometric singular integrals. Constructive and Computational Fractional Analysis recently is more and more in the center of mathematics because of their great applications in the real world. In this book, all presented is original work by the author given at a very general level to cover a maximum number of cases in various applications. The author applies generalized fractional differentiation techniques of Riemann-Liouville, Caputo and Canavati types and of fractional variable order to various kinds of inequalities such as of Opial, Hardy, Hilbert-Pachpatte and on the spherical shell. He continues with E. R. Love left- and right-side fractional integral inequalities. They follow fractional Landau inequalities, of left and right sides, univariate and multivariate, including ones for Semigroups. These are developed to all possible directions, and right-side multivariate fractional Taylor formulae are proven for the purpose. It continues with several Gronwall fractional inequalities of variable order. This book results are expected to find applications in many areas of pure and applied mathematics. As such this book is suitable for researchers, graduate students and seminars of the above disciplines, also to be in all science and engineering libraries.
This book discusses the interplay between statistics, data science, machine learning and artificial intelligence, with a focus on environmental science, the natural sciences, and technology. It covers the state of the art from both a theoretical and a practical viewpoint and describes how to successfully apply machine learning methods, demonstrating the benefits of statistics for modeling and analyzing high-dimensional and big data. The book's expert contributions include theoretical studies of machine learning methods, expositions of general methodologies for sound statistical analyses of data as well as novel approaches to modeling and analyzing data for specific problems and areas. In terms of applications, the contributions deal with data as arising in industrial quality control, autonomous driving, transportation and traffic, chip manufacturing, photovoltaics, football, transmission of infectious diseases, Covid-19 and public health. The book will appeal to statisticians and data scientists, as well as engineers and computer scientists working in related fields or applications.
Representation learning in heterogeneous graphs (HG) is intended to provide a meaningful vector representation for each node so as to facilitate downstream applications such as link prediction, personalized recommendation, node classification, etc. This task, however, is challenging not only because of the need to incorporate heterogeneous structural (graph) information consisting of multiple types of node and edge, but also the need to consider heterogeneous attributes or types of content (e.g. text or image) associated with each node. Although considerable advances have been made in homogeneous (and heterogeneous) graph embedding, attributed graph embedding and graph neural networks, few are capable of simultaneously and effectively taking into account heterogeneous structural (graph) information as well as the heterogeneous content information of each node. In this book, we provide a comprehensive survey of current developments in HG representation learning. More importantly, we present the state-of-the-art in this field, including theoretical models and real applications that have been showcased at the top conferences and journals, such as TKDE, KDD, WWW, IJCAI and AAAI. The book has two major objectives: (1) to provide researchers with an understanding of the fundamental issues and a good point of departure for working in this rapidly expanding field, and (2) to present the latest research on applying heterogeneous graphs to model real systems and learning structural features of interaction systems. To the best of our knowledge, it is the first book to summarize the latest developments and present cutting-edge research on heterogeneous graph representation learning. To gain the most from it, readers should have a basic grasp of computer science, data mining and machine learning.
This book chronicles a 10-year introduction of blended learning into the delivery at a leading technological university, with a longstanding tradition of technology-enabled teaching and learning, and state-of-the-art infrastructure. Hence, both teachers and students were familiar with the idea of online courses. Despite this, the longitudinal experiment did not proceed as expected. Though few technical problems, it required behavioural changes from teachers and learners, thus unearthing a host of socio-technical issues, challenges, and conundrums. With the undercurrent of design ideals such as "tech for good", any industrial sector must examine whether digital platforms are credible substitutes or at best complementary. In this era of Industry 4.0, higher education, like any other industry, should not be about the creative destruction of what we value in universities, but their digital transformation. The book concludes with an agenda for large, repeatable Randomised Controlled Trials (RCTs) to validate digital platforms that could fulfil the aspirations of the key stakeholder groups - students, faculty, and regulators as well as delving into the role of Massive Open Online Courses (MOOCs) as surrogates for "fees-free" higher education and whether the design of such a HiEd 4.0 platform is even a credible proposition. Specifically, the book examines the data-driven evidence within a design-based research methodology to present outcomes of two alternative instructional designs evaluated - traditional lecturing and blended learning. Based on the research findings and statistical analysis, it concludes that the inexorable shift to online delivery of education must be guided by informed educational management and innovation.
bridges ML and Optimisation; discusses optimisation techniques to improve ML algorithms for big data problems; identifies key research areas to solve large-scale machine learning problems; identifies recent research directions to solve major areas to tackle the challenge
RDF-based knowledge graphs require additional formalisms to be fully context-aware, which is presented in this book. This book also provides a collection of provenance techniques and state-of-the-art metadata-enhanced, provenance-aware, knowledge graph-based representations across multiple application domains, in order to demonstrate how to combine graph-based data models and provenance representations. This is important to make statements authoritative, verifiable, and reproducible, such as in biomedical, pharmaceutical, and cybersecurity applications, where the data source and generator can be just as important as the data itself. Capturing provenance is critical to ensure sound experimental results and rigorously designed research studies for patient and drug safety, pathology reports, and medical evidence generation. Similarly, provenance is needed for cyberthreat intelligence dashboards and attack maps that aggregate and/or fuse heterogeneous data from disparate data sources to differentiate between unimportant online events and dangerous cyberattacks, which is demonstrated in this book. Without provenance, data reliability and trustworthiness might be limited, causing data reuse, trust, reproducibility and accountability issues. This book primarily targets researchers who utilize knowledge graphs in their methods and approaches (this includes researchers from a variety of domains, such as cybersecurity, eHealth, data science, Semantic Web, etc.). This book collects core facts for the state of the art in provenance approaches and techniques, complemented by a critical review of existing approaches. New research directions are also provided that combine data science and knowledge graphs, for an increasingly important research topic.
This book bridges theoretical computer science and machine learning by exploring what the two sides can teach each other. It emphasizes the need for flexible, tractable models that better capture not what makes machine learning hard, but what makes it easy. Theoretical computer scientists will be introduced to important models in machine learning and to the main questions within the field. Machine learning researchers will be introduced to cutting-edge research in an accessible format, and gain familiarity with a modern, algorithmic toolkit, including the method of moments, tensor decompositions and convex programming relaxations. The treatment beyond worst-case analysis is to build a rigorous understanding about the approaches used in practice and to facilitate the discovery of exciting, new ways to solve important long-standing problems.
This book features high-quality, peer-reviewed research papers presented at the Fourth International Conference on Smart Technologies in Data Science and Communication (SMART-DSC 2021), held in Koneru Lakshmaiah Education Foundation, Guntur, Andhra Pradesh, India, on 18-19 February 2021. It includes innovative and novel contributions in the areas of data analytics, communication, and soft computing.
Deep connections exist between harmonic and applied analysis and the diverse yet connected topics of machine learning, data analysis, and imaging science. This volume explores these rapidly growing areas and features contributions presented at the second and third editions of the Summer Schools on Applied Harmonic Analysis, held at the University of Genova in 2017 and 2019. Each chapter offers an introduction to essential material and then demonstrates connections to more advanced research, with the aim of providing an accessible entrance for students and researchers. Topics covered include ill-posed problems; concentration inequalities; regularization and large-scale machine learning; unitarization of the radon transform on symmetric spaces; and proximal gradient methods for machine learning and imaging.
The technique of randomization has been employed to solve numerous prob lems of computing both sequentially and in parallel. Examples of randomized algorithms that are asymptotically better than their deterministic counterparts in solving various fundamental problems abound. Randomized algorithms have the advantages of simplicity and better performance both in theory and often in practice. This book is a collection of articles written by renowned experts in the area of randomized parallel computing. A brief introduction to randomized algorithms In the aflalysis of algorithms, at least three different measures of performance can be used: the best case, the worst case, and the average case. Often, the average case run time of an algorithm is much smaller than the worst case. 2 For instance, the worst case run time of Hoare's quicksort is O(n ), whereas its average case run time is only O( n log n). The average case analysis is conducted with an assumption on the input space. The assumption made to arrive at the O( n log n) average run time for quicksort is that each input permutation is equally likely. Clearly, any average case analysis is only as good as how valid the assumption made on the input space is. Randomized algorithms achieve superior performances without making any assumptions on the inputs by making coin flips within the algorithm. Any analysis done of randomized algorithms will be valid for all p0: .sible inputs."
This book features research papers presented at the International Conference on Emerging Technologies in Data Mining and Information Security (IEMIS 2020) held at the University of Engineering & Management, Kolkata, India, during July 2020. The book is organized in three volumes and includes high-quality research work by academicians and industrial experts in the field of computing and communication, including full-length papers, research-in-progress papers, and case studies related to all the areas of data mining, machine learning, Internet of things (IoT), and information security.
This book focuses on how the BOXES Methodology, which is based on the work of Donald Michie, is applied to ill-defined real-time control systems with minimal a priori knowledge of the system. The method is applied to a variety of systems including the familiar pole and cart. This second edition includes a new section that covers some further observations and thoughts, problems, and evolutionary extensions that the reader will find useful in their own implementation of the method. This second edition includes a new section on how to handle jittering about a system boundary which in turn causes replicated run times to become part of the learning mechanism. It also addresses the aging of data values using a forgetfulness factor that causes wrong values of merit to be calculated. Another question that is addressed is "Should a BOXES cell ever be considered fully trained and, if so, excluded from further dynamic updates". Finally, it expands on how system boundaries may be shifted using data from many runs using an evolutionary paradigm.
This textbook introduces major topics that include quantum bits, superposition, entanglement, logic gates, quantum search algorithm, quantum Fourier transform, inverse quantum Fourier transform, Shor's order-finding algorithm and phase estimation. Everyone can write algorithms and programs in the cloud making using IBM's quantum computers that support IBM Q Experience which contains the composer, open quantum assembly language, simulators and real quantum devices. Furthermore, this book teaches you how to use open quantum assembly language to write quantum programs for dealing with complex problems. Through numerous examples and exercises, readers will learn how to write a quantum program with open quantum assembly language for solving any problem from start to complete. This book includes six main chapters: *Quantum Bits and Quantum Gates-learn what quantum bits are, how to declare and measure them, what quantum gates are and how they work on a simulator or a real device in the cloud. *Boolean Algebra and its Applications-learn how to decompose CCNOT gate into six CNOT gates and nine gates of one bit and how to use NOT gates, CNOT gates and CCNOT gates to implement logic operations including NOT, OR, AND, NOR, NAND, Exclusive-OR (XOR) and Exclusive-NOR (XNOR). *Quantum Search Algorithm and its Applications-learn core concepts of quantum search algorithm and how to write quantum programs to implement core concepts of quantum search algorithm for solving two famous NP-complete problems that are the satisfiability problem in n Boolean variables and m clauses and the clique problem in a graph with n vertices and q edges. *Quantum Fourier Transform and its Applications-learn core concepts of quantum Fourier transform and inverse quantum Fourier transform and how to write quantum programs to implement them for solving two real applications that are to compute the period and the frequency of two given oracular functions. *Order-Finding and Factoring-learn core concepts of Shor's order-finding algorithm and how to write quantum programs to implement Shor's order-finding algorithm for completing the prime factorization to 15. Phase Estimation and its Applications-learn core concepts of phase estimation and quantum counting and how to write quantum programs to implement them to compute the number of solution(s) in the independent set problem in a graph with two vertices and one edge.
This book provides an overview of state-of-the-art implementations of quantum random number generators (QRNGs), and especially examines their relation to classical statistical randomness models and numerical techniques for computing random numbers. The reader - who ideally has a background in classical statistics, computer science, or cryptography - is introduced to the world of quantum bits step by step, and explicit relations between QRNGs and their classical counterparts are identified along the way. Random number generation is a major pillar of cryptography. Capitalizing on the randomness inherent in quantum phenomena is a rapidly evolving branch of quantum cryptography with countless applications for the future. The value of quantum randomness for cryptographic purposes is empirically demonstrated in statistical evaluations of QRNGs' performance compared to classical techniques for true and pseudorandom number generation. The book then provides an overview of technical implementations of QRNGs, before a concluding discussion of major achievements and remaining obstacles in the field rounds out the coverage, while also opening the door for future research directions.
This book describes how we can design and make efficient processors for high-performance computing, AI, and data science. Although there are many textbooks on the design of processors we do not have a widely accepted definition of the efficiency of a general-purpose computer architecture. Without a definition of the efficiency, it is difficult to make scientific approach to the processor design. In this book, a clear definition of efficiency is given and thus a scientific approach for processor design is made possible. In chapter 2, the history of the development of high-performance processor is overviewed, to discuss what quantity we can use to measure the efficiency of these processors. The proposed quantity is the ratio between the minimum possible energy consumption and the actual energy consumption for a given application using a given semiconductor technology. In chapter 3, whether or not this quantity can be used in practice is discussed, for many real-world applications. In chapter 4, general-purpose processors in the past and present are discussed from this viewpoint. In chapter 5, how we can actually design processors with near-optimal efficiencies is described, and in chapter 6 how we can program such processors. This book gives a new way to look at the field of the design of high-performance processors.
Reversible grammar allows computational models to be built that are equally well suited for the analysis and generation of natural language utterances. This task can be viewed from very different perspectives by theoretical and computational linguists, and computer scientists. The papers in this volume present a broad range of approaches to reversible, bi-directional, and non-directional grammar systems that have emerged in recent years. This is also the first collection entirely devoted to the problems of reversibility in natural language processing. Most papers collected in this volume are derived from presentations at a workshop held at the University of California at Berkeley in the summer of 1991 organised under the auspices of the Association for Computational Linguistics. This book will be a valuable reference to researchers in linguistics and computer science with interests in computational linguistics, natural language processing, and machine translation, as well as in practical aspects of computability.
This book discusses one of the major applications of artificial intelligence: the use of machine learning to extract useful information from multimodal data. It discusses the optimization methods that help minimize the error in developing patterns and classifications, which further helps improve prediction and decision-making. The book also presents formulations of real-world machine learning problems, and discusses AI solution methodologies as standalone or hybrid approaches. Lastly, it proposes novel metaheuristic methods to solve complex machine learning problems. Featuring valuable insights, the book helps readers explore new avenues leading toward multidisciplinary research discussions.
An invaluable resource for working programmers, as well as a fount of useful algorithmic tools for computer scientists, astronomers, and other calendar enthusiasts, The Ultimate Edition updates and expands the previous edition to achieve more accurate results and present new calendar variants. The book now includes coverage of Unix dates, Italian time, the Akan, Icelandic, Saudi Arabian Umm al-Qura, and Babylonian calendars. There are also expanded treatments of the observational Islamic and Hebrew calendars and brief discussions of the Samaritan and Nepalese calendars. Several of the astronomical functions have been rewritten to produce more accurate results and to include calculations of moonrise and moonset. The authors frame the calendars of the world in a completely algorithmic form, allowing easy conversion among these calendars and the determination of secular and religious holidays. LISP code for all the algorithms is available in machine-readable form.
Techniques and principles of minimax theory play a key role in many areas of research, including game theory, optimization, and computational complexity. In general, a minimax problem can be formulated as min max f(x, y) (1) ",EX !lEY where f(x, y) is a function defined on the product of X and Y spaces. There are two basic issues regarding minimax problems: The first issue concerns the establishment of sufficient and necessary conditions for equality minmaxf(x,y) = maxminf(x,y). (2) "'EX !lEY !lEY "'EX The classical minimax theorem of von Neumann is a result of this type. Duality theory in linear and convex quadratic programming interprets minimax theory in a different way. The second issue concerns the establishment of sufficient and necessary conditions for values of the variables x and y that achieve the global minimax function value f(x*, y*) = minmaxf(x, y). (3) "'EX !lEY There are two developments in minimax theory that we would like to mention. |
You may like...
Evolutionary Computation and Complex…
Jing Liu, Hussein A. Abbass, …
Hardcover
R2,653
Discovery Miles 26 530
Principles of Radio Navigation for…
Sauta O.I., Shatrakov A.Y., …
Hardcover
R2,653
Discovery Miles 26 530
Computational and Statistical Methods…
Shen Liu, James McGree, …
Hardcover
R1,802
Discovery Miles 18 020
|