![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > General theory of computing > Data structures
The 14 contributed chapters in this book survey the most recent developments in high-performance algorithms for NGS data, offering fundamental insights and technical information specifically on indexing, compression and storage; error correction; alignment; and assembly. The book will be of value to researchers, practitioners and students engaged with bioinformatics, computer science, mathematics, statistics and life sciences.
This book focuses on new and emerging data mining solutions that offer a greater level of transparency than existing solutions. Transparent data mining solutions with desirable properties (e.g. effective, fully automatic, scalable) are covered in the book. Experimental findings of transparent solutions are tailored to different domain experts, and experimental metrics for evaluating algorithmic transparency are presented. The book also discusses societal effects of black box vs. transparent approaches to data mining, as well as real-world use cases for these approaches.As algorithms increasingly support different aspects of modern life, a greater level of transparency is sorely needed, not least because discrimination and biases have to be avoided. With contributions from domain experts, this book provides an overview of an emerging area of data mining that has profound societal consequences, and provides the technical background to for readers to contribute to the field or to put existing approaches to practical use.
hebookpresentedtothereaderisdevotedtotime-dependentscheduling. TScheduling problems, in general, consist in the allocation of resources over time in order to perform a set of jobs. Any allocation that meets all requirements concerning the jobs and resources is called a feasible schedule. The quality of a schedule is measured by a criterion function. The aim of scheduling is to ?nd, among all feasible schedules, a schedule that optimizes the criterion function. A solution to an arbitrary scheduling problem consists in giving a polynomial-time algorithm generating either an optimal schedule or a schedule that is close to the optimal one, if the given scheduling problem has been proved to be computationally intractable. The scheduling problems are subject of interest of the scheduling theory, originated in mid-?fties of the twentieth century. The theory has been developing dynamically and new research areas constantly come into existence. The subject of this book, ti- dependent scheduling, is one of such areas. In time-dependent scheduling, the processing time of a job is variable and depends on the starting time of the job. This crucial assumption allows us to apply the scheduling theory to a broader spectrum of problems. For example, in the framework of the time-dependent scheduling theory we may consider the problems of repayment of multiple loans, ?re ?ghting and maintenance assignments. In this book, we will discuss algorithms and complexity issues concerning various time-dependent scheduling problems.
A collection of surveys and research papers on mathematical software and algorithms. The common thread is that the field of mathematical applications lies on the border between algebra and geometry. Topics include polyhedral geometry, elimination theory, algebraic surfaces, Gröbner bases, triangulations of point sets and the mutual relationship. This diversity is accompanied by the abundance of available software systems which often handle only special mathematical aspects. This is why the volume also focuses on solutions to the integration of mathematical software systems. This includes low-level and XML based high-level communication channels as well as general frameworks for modular systems.
This thesis introduces novel and significant results regarding the analysis and synthesis of positive systems, especially under l1 and L1 performance. It describes stability analysis, controller synthesis, and bounding positivity-preserving observer and filtering design for a variety of both discrete and continuous positive systems. It subsequently derives computationally efficient solutions based on linear programming in terms of matrix inequalities, as well as a number of analytical solutions obtained for special cases. The thesis applies a range of novel approaches and fundamental techniques to the further study of positive systems, thus contributing significantly to the theory of positive systems, a "hot topic" in the field of control.
This book contains some selected papers from the International Conference on Extreme Learning Machine (ELM) 2017, held in Yantai, China, October 4-7, 2017. The book covers theories, algorithms and applications of ELM. Extreme Learning Machines (ELM) aims to enable pervasive learning and pervasive intelligence. As advocated by ELM theories, it is exciting to see the convergence of machine learning and biological learning from the long-term point of view. ELM may be one of the fundamental `learning particles' filling the gaps between machine learning and biological learning (of which activation functions are even unknown). ELM represents a suite of (machine and biological) learning techniques in which hidden neurons need not be tuned: inherited from their ancestors or randomly generated. ELM learning theories show that effective learning algorithms can be derived based on randomly generated hidden neurons (biological neurons, artificial neurons, wavelets, Fourier series, etc) as long as they are nonlinear piecewise continuous, independent of training data and application environments. Increasingly, evidence from neuroscience suggests that similar principles apply in biological learning systems. ELM theories and algorithms argue that "random hidden neurons" capture an essential aspect of biological learning mechanisms as well as the intuitive sense that the efficiency of biological learning need not rely on computing power of neurons. ELM theories thus hint at possible reasons why the brain is more intelligent and effective than current computers. This conference will provide a forum for academics, researchers and engineers to share and exchange R&D experience on both theoretical studies and practical applications of the ELM technique and brain learning. It gives readers a glance of the most recent advances of ELM.
Thisvolumecontainstheinvitedandregularpaperspresentedat TCS 2010, the 6thIFIP International Conference on Theoretical Computer Science, organised by IFIP Tech- cal Committee 1 (Foundations of Computer Science) and IFIP WG 2.2 (Formal - scriptions of Programming Concepts) in association with SIGACT and EATCS. TCS 2010 was part of the World Computer Congress held in Brisbane, Australia, during September 20-23, 2010 ( ). TCS 2010 is composed of two main areas: (A) Algorithms, Complexity and Models of Computation, and (B) Logic, Semantics, Speci?cation and Veri?cation. The selection process led to the acceptance of 23 papers out of 39 submissions, eachofwhichwasreviewedbythreeProgrammeCommitteemembers.TheProgramme Committee discussion was held electronically using Easychair. The invited speakers at TCS 2010 are: Rob van Glabbeek (NICTA, Australia) Bart Jacobs (Nijmegen, The Netherlands) Catuscia Palamidessi (INRIA and LIX, Paris, France) Sabina Rossi (Venice, Italy) James Harland (Australia) and Barry Jay (Australia) acted as TCS 2010 Chairs. We take this occasion to thank the members of the Programme Committees and the external reviewers for the professional and timely work; the conference Chairs for their support; the invited speakers for their scholarly contribution; and of course the authors for submitting their work to TCS 2010
Graph theory is a specific concept that has numerous applications throughout many industries. Despite the advancement of this technique, graph theory can still yield ambiguous and imprecise results. In order to cut down on these indeterminate factors, neutrosophic logic has emerged as an applicable solution that is gaining significant attention in solving many real-life decision-making problems that involve uncertainty, impreciseness, vagueness, incompleteness, inconsistency, and indeterminacy. However, empirical research on this specific graph set is lacking. Neutrosophic Graph Theory and Algorithms is a collection of innovative research on the methods and applications of neutrosophic sets and logic within various fields including systems analysis, economics, and transportation. While highlighting topics including linear programming, decision-making methods, and homomorphism, this book is ideally designed for programmers, researchers, data scientists, mathematicians, designers, educators, researchers, academicians, and students seeking current research on the various methods and applications of graph theory.
This book constitutes the refereed proceedings of the 27th IFIP TC 11 International Information Security Conference, SEC 2012, held in Heraklion, Crete, Greece, in June 2012. The 42 revised full papers presented together with 11 short papers were carefully reviewed and selected from 167 submissions. The papers are organized in topical sections on attacks and malicious code, security architectures, system security, access control, database security, privacy attitudes and properties, social networks and social engineering, applied cryptography, anonymity and trust, usable security, security and trust models, security economics, and authentication and delegation.
The book is a collection of invited papers on Computational Intelligence for Privacy and Security. The majority of the chapters are extended versions of works presented at the special session on Computational Intelligence for Privacy and Security of the International Joint Conference on Neural Networks (IJCNN-2010) held July 2010 in Barcelona, Spain. The book is devoted to Computational Intelligence for Privacy and Security. It provides an overview of the most recent advances on the Computational Intelligence techniques being developed for Privacy and Security. The book will be of interest to researchers in industry and academics and to post-graduate students interested in the latest advances and developments in the field of Computational Intelligence for Privacy and Security.
Data mining is a very active research area with many successful real-world app- cations. It consists of a set of concepts and methods used to extract interesting or useful knowledge (or patterns) from real-world datasets, providing valuable support for decision making in industry, business, government, and science. Although there are already many types of data mining algorithms available in the literature, it is still dif cult for users to choose the best possible data mining algorithm for their particular data mining problem. In addition, data mining al- rithms have been manually designed; therefore they incorporate human biases and preferences. This book proposes a new approach to the design of data mining algorithms. - stead of relying on the slow and ad hoc process of manual algorithm design, this book proposes systematically automating the design of data mining algorithms with an evolutionary computation approach. More precisely, we propose a genetic p- gramming system (a type of evolutionary computation method that evolves c- puter programs) to automate the design of rule induction algorithms, a type of cl- si cation method that discovers a set of classi cation rules from data. We focus on genetic programming in this book because it is the paradigmatic type of machine learning method for automating the generation of programs and because it has the advantage of performing a global search in the space of candidate solutions (data mining algorithms in our case), but in principle other types of search methods for this task could be investigated in the future.
In operations research and computer science it is common practice to evaluate the performance of optimization algorithms on the basis of computational results, and the experimental approach should follow accepted principles that guarantee the reliability and reproducibility of results. However, computational experiments differ from those in other sciences, and the last decade has seen considerable methodological research devoted to understanding the particular features of such experiments and assessing the related statistical methods. This book consists of methodological contributions on different scenarios of experimental analysis. The first part overviews the main issues in the experimental analysis of algorithms, and discusses the experimental cycle of algorithm development; the second part treats the characterization by means of statistical distributions of algorithm performance in terms of solution quality, runtime and other measures; and the third part collects advanced methods from experimental design for configuring and tuning algorithms on a specific class of instances with the goal of using the least amount of experimentation. The contributor list includes leading scientists in algorithm design, statistical design, optimization and heuristics, and most chapters provide theoretical background and are enriched with case studies. This book is written for researchers and practitioners in operations research and computer science who wish to improve the experimental assessment of optimization algorithms and, consequently, their design.
This book investigates the susceptibility of intrinsic physically unclonable function (PUF) implementations on reconfigurable hardware to optical semi-invasive attacks from the chip backside. It explores different classes of optical attacks, particularly photonic emission analysis, laser fault injection, and optical contactless probing. By applying these techniques, the book demonstrates that the secrets generated by a PUF can be predicted, manipulated or directly probed without affecting the behavior of the PUF. It subsequently discusses the cost and feasibility of launching such attacks against the very latest hardware technologies in a real scenario. The author discusses why PUFs are not tamper-evident in their current configuration, and therefore, PUFs alone cannot raise the security level of key storage. The author then reviews the potential and already implemented countermeasures, which can remedy PUFs' security-related shortcomings and make them resistant to optical side-channel and optical fault attacks. Lastly, by making selected modifications to the functionality of an existing PUF architecture, the book presents a prototype tamper-evident sensor for detecting optical contactless probing attempts.
The authors give a detailed summary about the fundamentals and the historical background of digital communication. This includes an overview of the encoding principles and algorithms of textual information, audio information, as well as images, graphics, and video in the Internet. Furthermore the fundamentals of computer networking, digital security and cryptography are covered. Thus, the book provides a well-founded access to communication technology of computer networks, the internet and the WWW. Numerous pictures and images, a subject-index and a detailed list of historical personalities including a glossary for each chapter increase the practical benefit of this book that is well suited as well as for undergraduate students as for working practitioners.
Articles in this book examine various materials and how to determine directly the limit state of a structure, in the sense of limit analysis and shakedown analysis. Apart from classical applications in mechanical and civil engineering contexts, the book reports on the emerging field of material design beyond the elastic limit, which has further industrial design and technological applications. Readers will discover that "Direct Methods" and the techniques presented here can in fact be used to numerically estimate the strength of structured materials such as composites or nano-materials, which represent fruitful fields of future applications. Leading researchers outline the latest computational tools and optimization techniques and explore the possibility of obtaining information on the limit state of a structure whose post-elastic loading path and constitutive behavior are not well defined or well known. Readers will discover how Direct Methods allow rapid and direct access to requested information in mathematically constructive manners without cumbersome step-by-step computation. Both researchers already interested or involved in the field and practical engineers who want to have a panorama of modern methods for structural safety assessment will find this book valuable. It provides the reader with the latest developments and a significant amount of references on the topic.
The book offers an original view on channel coding, based on a unitary approach to block and convolutional codes for error correction. It presents both new concepts and new families of codes. For example, lengthened and modified lengthened cyclic codes are introduced as a bridge towards time-invariant convolutional codes and their extension to time-varying versions. The novel families of codes include turbo codes and low-density parity check (LDPC) codes, the features of which are justified from the structural properties of the component codes. Design procedures for regular LDPC codes are proposed, supported by the presented theory. Quasi-cyclic LDPC codes, in block or convolutional form, represent one of the most original contributions of the book. The use of more than 100 examples allows the reader gradually to gain an understanding of the theory, and the provision of a list of more than 150 definitions, indexed at the end of the book, permits rapid location of sought information.
Graph algorithms is a well-established subject in mathematics and computer science. Beyond classical application fields, such as approximation, combinatorial optimization, graphics, and operations research, graph algorithms have recently attracted increased attention from computational molecular biology and computational chemistry. Centered around the fundamental issue of graph isomorphism, this text goes beyond classical graph problems of shortest paths, spanning trees, flows in networks, and matchings in bipartite graphs. Advanced algorithmic results and techniques of practical relevance are presented in a coherent and consolidated way. This book introduces graph algorithms on an intuitive basis followed by a detailed exposition in a literate programming style, with correctness proofs as well as worst-case analyses. Furthermore, full C++ implementations of all algorithms presented are given using the LEDA library of efficient data structures and algorithms.
Mixed-Signal Embedded Microcontrollers are commonly used in integrating analog components needed to control non-digital electronic systems. They are used in automatically controlled devices and products, such as automobile engine control systems, wireless remote controllers, office machines, home appliances, power tools, and toys. Microcontrollers make it economical to digitally control even more devices and processes by reducing the size and cost, compared to a design that uses a separate microprocessor, memory, and input/output devices. In many undergraduate and post-graduate courses, teaching of mixed-signal microcontrollers and their use for project work has become compulsory. Students face a lot of difficulties when they have to interface a microcontroller with the electronics they deal with. This book addresses some issues of interfacing the microcontrollers and describes some project implementations with the Silicon Lab C8051F020 mixed-signal microcontroller. The intended readers are college and university students specializing in electronics, computer systems engineering, electrical and electronics engineering; researchers involved with electronics based system, practitioners, technicians and in general anybody interested in microcontrollers based projects.
This book provides the most updated information of how membrane lipids mediate protein signaling from studies carried out in animal and plant cells. Also, there are some chapters that go beyond and expand these studies of protein-lipid interactions at the structural level. The book begins with a literature review from investigations associated to sphingolipids, followed by studies that describe the role of phosphoinositides in signaling and closing with the function of other key lipids in signaling at the plasma membrane and intracellular organelles.
"Optimal Design of Distributed Control and Embedded Systems "focuses on the design of special control and scheduling algorithms based on system structural properties as well as on analysis of the influence of induced time-delay on systems performances. It treats the optimal design of distributed and embedded control systems (DCESs) with respect to communication and calculation-resource constraints, quantization aspects, and potential time-delays induced by the associated communication and calculation model. Particular emphasis is put on optimal control signal scheduling based on the system state. In order to render this complex optimization problem feasible in real time, a time decomposition is based on periodicity induced by the static scheduling is operated. The authors present a co-design approach which subsumes the synthesis of the optimal control laws and the generation of an optimal schedule of control signals on real-time networks as well as the execution of control tasks on a single processor. The authors also operate a control structure modification or a control switching based on a thorough analysis of the influence of the induced time-delay system influence on stability and system performance in order to optimize DCES performance in case of calculation and communication resource limitations. Although the richness and variety of classes of DCES preclude a completely comprehensive treatment or a single best method of approaching them all, this co-design approach has the best chance of rendering this problem feasible and finding the optimal or some sub-optimal solution. The text is rounded out with references to such applications as car suspension and unmanned vehicles. "Optimal Design of Distributed Control and Embedded Systems" will be of most interest to academic researchers working on the mathematical theory of DCES but the wide range of environments in which they are used also promotes the relevance of the text for control practitioners working in the avionics, automotive, energy-production, space exploration and many other industries."
This is an examination of the history and the state of the art of the quest for visualizing scientific knowledge and the dynamics of its development. Through an interdisciplinary perspective this book presents profound visions, pivotal advances, and insightful contributions made by generations of researchers and professionals, which portrays a holistic view of the underlying principles and mechanisms of the development of science. This updated and extended second edition: highlights the latest advances in mapping scientific frontiers examines the foundations of strategies, principles, and design patterns provides an integrated and holistic account of major developments across disciplinary boundaries "Anyone who tries to follow the exponential growth of the literature on citation analysis and scientometrics knows how difficult it is to keep pace. Chaomei Chen has identified the significant methods and applications in visual graphics and made them clear to the uninitiated. Derek Price would have loved this book which not only pays homage to him but also to the key players in information science and a wide variety of others in the sociology and history of science." - Eugene Garfield "This is a wide ranging book on information visualization, with a specific focus on science mapping. Science mapping is still in its infancy and many intellectual challenges remain to be investigated and many of which are outlined in the final chapter. In this new edition Chaomei Chen has provided an essential text, useful both as a primer for new entrants and as a comprehensive overview of recent developments for the seasoned practitioner." - Henry Small Chaomei Chen is a Professor in the College of Information Science and Technology at Drexel University, Philadelphia, USA, and a ChangJiang Scholar at Dalian University of Technology, Dalian, China. He is the Editor-in-Chief of Information Visualization and the author of Turning Points: The Nature of Creativity (Springer, 2012) and Information Visualization: Beyond the Horizon (Springer, 2004, 2006).
This new edition includes the latest advances and developments in computational probability involving A Probability Programming Language (APPL). The book examines and presents, in a systematic manner, computational probability methods that encompass data structures and algorithms. The developed techniques address problems that require exact probability calculations, many of which have been considered intractable in the past. The book addresses the plight of the probabilist by providing algorithms to perform calculations associated with random variables. Computational Probability: Algorithms and Applications in the Mathematical Sciences, 2nd Edition begins with an introductory chapter that contains short examples involving the elementary use of APPL. Chapter 2 reviews the Maple data structures and functions necessary to implement APPL. This is followed by a discussion of the development of the data structures and algorithms (Chapters 3-6 for continuous random variables and Chapters 7-9 for discrete random variables) used in APPL. The book concludes with Chapters 10-15 introducing a sampling of various applications in the mathematical sciences. This book should appeal to researchers in the mathematical sciences with an interest in applied probability and instructors using the book for a special topics course in computational probability taught in a mathematics, statistics, operations research, management science, or industrial engineering department.
This book develops a new approach called parameter advising for finding a parameter setting for a sequence aligner that yields a quality alignment of a given set of input sequences. In this framework, a parameter advisor is a procedure that automatically chooses a parameter setting for the input, and has two main ingredients: (a) the set of parameter choices considered by the advisor, and (b) an estimator of alignment accuracy used to rank alignments produced by the aligner. On coupling a parameter advisor with an aligner, once the advisor is trained in a learning phase, the user simply inputs sequences to align, and receives an output alignment from the aligner, where the advisor has automatically selected the parameter setting. The chapters first lay out the foundations of parameter advising, and then cover applications and extensions of advising. The content * examines formulations of parameter advising and their computational complexity, * develops methods for learning good accuracy estimators, * presents approximation algorithms for finding good sets of parameter choices, and * assesses software implementations of advising that perform well on real biological data. Also explored are applications of parameter advising to * adaptive local realignment, where advising is performed on local regions of the sequences to automatically adapt to varying mutation rates, and * ensemble alignment, where advising is applied to an ensemble of aligners to effectively yield a new aligner of higher quality than the individual aligners in the ensemble. The book concludes by offering future directions in advising research.
This timely review volume summarizes the state-of-the-art developments in nature-inspired algorithms and applications with the emphasis on swarm intelligence and bio-inspired computation. Topics include the analysis and overview of swarm intelligence and evolutionary computation, hybrid metaheuristic algorithms, bat algorithm, discrete cuckoo search, firefly algorithm, particle swarm optimization, and harmony search as well as convergent hybridization. Application case studies have focused on the dehydration of fruits and vegetables by the firefly algorithm and goal programming, feature selection by the binary flower pollination algorithm, job shop scheduling, single row facility layout optimization, training of feed-forward neural networks, damage and stiffness identification, synthesis of cross-ambiguity functions by the bat algorithm, web document clustering, truss analysis, water distribution networks, sustainable building designs and others. As a timely review, this book can serve as an ideal reference for graduates, lecturers, engineers and researchers in computer science, evolutionary computing, artificial intelligence, machine learning, computational intelligence, data mining, engineering optimization and designs.
Recently many researchers are working on cluster analysis as a main tool for exploratory data analysis and data mining. A notable feature is that specialists in di?erent ?elds of sciences are considering the tool of data clustering to be useful. A major reason is that clustering algorithms and software are ?exible in thesensethatdi?erentmathematicalframeworksareemployedinthealgorithms and a user can select a suitable method according to his application. Moreover clusteringalgorithmshavedi?erentoutputsrangingfromtheolddendrogramsof agglomerativeclustering to more recent self-organizingmaps. Thus, a researcher or user can choose an appropriate output suited to his purpose, which is another ?exibility of the methods of clustering. An old and still most popular method is the K-means which use K cluster centers. A group of data is gathered around a cluster center and thus forms a cluster. The main subject of this book is the fuzzy c-means proposed by Dunn and Bezdek and their variations including recent studies. A main reasonwhy we concentrate on fuzzy c-means is that most methodology and application studies infuzzy clusteringusefuzzy c-means, andfuzzy c-meansshouldbe consideredto beamajortechniqueofclusteringingeneral, regardlesswhetheroneisinterested in fuzzy methods or not. Moreover recent advances in clustering techniques are rapid and we requirea new textbook that includes recent algorithms.We should also note that several books have recently been published but the contents do not include some methods studied here |
You may like...
Concept Parsing Algorithms (CPA) for…
Uri Shafrir, Masha Etkind
Hardcover
R3,276
Discovery Miles 32 760
Air Traffic Control Automated Systems
Bestugin A.R., Eshenko A.A., …
Hardcover
R3,133
Discovery Miles 31 330
|