![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > General theory of computing > Mathematical theory of computation
This book offers a comprehensive collection of the most advanced numerical techniques for the efficient and effective solution of simulation and optimization problems governed by systems of time-dependent differential equations. The contributions present various approaches to time domain decomposition, focusing on multiple shooting and parareal algorithms. The range of topics covers theoretical analysis of the methods, as well as their algorithmic formulation and guidelines for practical implementation. Selected examples show that the discussed approaches are mandatory for the solution of challenging practical problems. The practicability and efficiency of the presented methods is illustrated by several case studies from fluid dynamics, data compression, image processing and computational biology, giving rise to possible new research topics. This volume, resulting from the workshop Multiple Shooting and Time Domain Decomposition Methods, held in Heidelberg in May 2013, will be of great interest to applied mathematicians, computer scientists and all scientists using mathematical methods.
The book opens with a short introduction to Indian music, in particular classical Hindustani music, followed by a chapter on the role of statistics in computational musicology. The authors then show how to analyze musical structure using Rubato, the music software package for statistical analysis, in particular addressing modeling, melodic similarity and lengths, and entropy analysis; they then show how to analyze musical performance. Finally, they explain how the concept of seminatural composition can help a music composer to obtain the opening line of a raga-based song using Monte Carlo simulation. The book will be of interest to musicians and musicologists, particularly those engaged with Indian music.
This book is devoted to the state-of-the-art in all aspects of fireworks algorithm (FWA), with particular emphasis on the efficient improved versions of FWA. It describes the most substantial theoretical analysis including basic principle and implementation of FWA and modeling and theoretical analysis of FWA. It covers exhaustively the key recent significant research into the improvements of FWA so far. In addition, the book describes a few advanced topics in the research of FWA, including multi-objective optimization (MOO), discrete FWA (DFWA) for combinatorial optimization, and GPU-based FWA for parallel implementation. In sequels, several successful applications of FWA on non-negative matrix factorization (NMF), text clustering, pattern recognition, and seismic inversion problem, and swarm robotics, are illustrated in details, which might shed new light on more real-world applications in future. Addressing a multidisciplinary topic, it will appeal to researchers and professionals in the areas of metahuristics, swarm intelligence, evolutionary computation, complex optimization solving, etc.
This book presents a collection of state of the art research on GPU Computing and Application. The major part of this book is selected from the work presented at the 2013 Symposium on GPU Computing and Applications held in Nanyang Technological University, Singapore (Oct 9, 2013). Three major domains of GPU application are covered in the book including (1) Engineering design and simulation; (2) Biomedical Sciences; and (3) Interactive & Digital Media. The book also addresses the fundamental issues in GPU computing with a focus on big data processing. Researchers and developers in GPU Computing and Applications will benefit from this book. Training professionals and educators can also benefit from this book to learn the possible application of GPU technology in various areas.
The book is based on the PhD thesis "Descriptive Set Theoretic Methods in Automata Theory," awarded the E.W. Beth Prize in 2015 for outstanding dissertations in the fields of logic, language, and information. The thesis reveals unexpected connections between advanced concepts in logic, descriptive set theory, topology, and automata theory and provides many deep insights into the interplay between these fields. It opens new perspectives on central problems in the theory of automata on infinite words and trees and offers very impressive advances in this theory from the point of view of topology. "...the thesis of Michal Skrzypczak offers certainly what we expect from excellent mathematics: new unexpected connections between a priori distinct concepts, and proofs involving enlightening ideas." Thomas Colcombet.
Edited in collaboration with FoLLI, the Association of Logic, Language and Information this book constitutes the refereed proceedings of the 23rd Workshop on Logic, Language, Information and Communication, WoLLIC 2016, held in Puebla, Mexico, in August 2016.The 23 contributed papers, presented together with 9 invited lectures and tutorials, were carefully reviewed and selected from 33 submissions. The focus of the workshop is to provide a forum on inter-disciplinary research involving formal logic, computing and programming theory, and natural language and reasoning.
Walter Gautschi has written extensively on topics ranging from special functions, quadrature and orthogonal polynomials to difference and differential equations, software implementations, and the history of mathematics. He is world renowned for his pioneering work in numerical analysis and constructive orthogonal polynomials, including a definitive textbook in the former, and a monograph in the latter area. This three-volume set, Walter Gautschi: Selected Works with Commentaries, is a compilation of Gautschi's most influential papers and includes commentaries by leading experts. The work begins with a detailed biographical section and ends with a section commemorating Walter's prematurely deceased twin brother. This title will appeal to graduate students and researchers in numerical analysis, as well as to historians of science. Selected Works with Commentaries, Vol. 1 Numerical Conditioning Special Functions Interpolation and Approximation Selected Works with Commentaries, Vol. 2 Orthogonal Polynomials on the Real Line Orthogonal Polynomials on the Semicircle Chebyshev Quadrature Kronrod and Other Quadratures Gauss-type Quadrature Selected Works with Commentaries, Vol. 3 Linear Difference Equations Ordinary Differential Equations Software History and Biography Miscellanea Works of Werner Gautschi
Providing a wide variety of technologies for ensuring the safety and dependability of cyber-physical systems (CPS), this book offers a comprehensive introduction to the architecture-centric modeling, analysis, and verification of CPS. In particular, it focuses on model driven engineering methods including architecture description languages, virtual prototyping, and formal analysis methods. CPS are based on a new design paradigm intended to enable emerging software-intensive systems. Embedded computers and networks monitor and control the physical processes, usually with the help of feedback loops where physical processes affect computations and vice versa. The principal challenges in system design lie in this constant interaction of software, hardware and physics. Developing reliable CPS has become a critical issue for the industry and society, because many applications such as transportation, power distribution, medical equipment and tele-medicine are dependent on CPS. Safety and security requirements must be ensured by means of powerful validation tools. Satisfying such requirements, including quality of service, implies having formally proven the required properties of the system before it is deployed. The book is concerned with internationally standardized modeling languages such as AADL, SysML, and MARTE. As the effectiveness of the technologies is demonstrated with industrial sample cases from the automotive and aerospace sectors, links between the methods presented and industrial problems are clearly understandable. Each chapter is self-contained, addressing specific scientific or engineering problems, and identifying further issues. In closing, it includes perspectives on future directions in CPS design from an architecture analysis viewpoint.
This book questions the relevance of computation to the physical universe. Our theories deliver computational descriptions, but the gaps and discontinuities in our grasp suggest a need for continued discourse between researchers from different disciplines, and this book is unique in its focus on the mathematical theory of incomputability and its relevance for the real world. The core of the book consists of thirteen chapters in five parts on extended models of computation; the search for natural examples of incomputable objects; mind, matter, and computation; the nature of information, complexity, and randomness; and the mathematics of emergence and morphogenesis. This book will be of interest to researchers in the areas of theoretical computer science, mathematical logic, and philosophy.
This book aims at gathering roboticists, control theorists, neuroscientists, and mathematicians, in order to promote a multidisciplinary research on movement analysis. It follows the workshop " Geometric and Numerical Foundations of Movements " held at LAAS-CNRS in Toulouse in November 2015[1]. Its objective is to lay the foundations for a mutual understanding that is essential for synergetic development in motion research. In particular, the book promotes applications to robotics --and control in general-- of new optimization techniques based on recent results from real algebraic geometry.
This book constitutes revised selected papers from the 5th International Conference on Operations Research and Enterprise Systems, ICORES 2016, held in Rome, Italy, in February 2016. The 14 papers presented in this volume were carefully reviewed and selection from a total of 75 submissions. They are organized in topical sections named: methodologies and technologies; and applications.
This book comprises nine selected works on numerical and computational methods for solving multiobjective optimization, game theory, and machine learning problems. It provides extended versions of selected papers from various fields of science such as computer science, mathematics and engineering that were presented at EVOLVE 2013 held in July 2013 at Leiden University in the Netherlands. The internationally peer-reviewed papers include original work on important topics in both theory and applications, such as the role of diversity in optimization, statistical approaches to combinatorial optimization, computational game theory, and cell mapping techniques for numerical landscape exploration. Applications focus on aspects including robustness, handling multiple objectives, and complex search spaces in engineering design and computational biology.
This book constitutes the refereed proceedings of the 15th International Scientific Conference on Information Technologies and Mathematical Modeling, named after A. F. Terpugov, ITMM 2016, held in Katun, Russia, in September 2016. The 33 full papers presented together with 4 short papers were carefully reviewed and selected from 96 submissions. They are devoted to new results in the queueing theory and its applications, addressing specialists in probability theory, random processes, mathematical modeling as well as engineers dealing with logical and technical design and operational management of telecommunication and computer networks.
This book constitutes the refereed proceedings of the 23rd International Static Analysis Symposium, SAS 2016, held in Edinburgh, UK, in September 2016. The 21 papers presented in this volume were carefully reviewed and selected from 55 submissions. The contributions cover a variety of multi-disciplinary topics in abstract domains; abstract interpretation; abstract testing; bug detection; data flow analysis; model checking; new applications; program transformation; program verification; security analysis; theoretical frameworks; and type checking.
Among the various multi-level formulations of mathematical models in decision making processes, this book focuses on the bi-level model. Being the most frequently used, the bi-level model addresses conflicts which exist in multi-level decision making processes. From the perspective of bi-level structure and uncertainty, this book takes real-life problems as the background, focuses on the so-called random-like uncertainty, and develops the general framework of random-like bi-level decision making problems. The random-like uncertainty considered in this book includes random phenomenon, random-overlapped random (Ra-Ra) phenomenon and fuzzy-overlapped random (Ra-Fu) phenomenon. Basic theory, models, algorithms and practical applications for different types of random-like bi-level decision making problems are also presented in this book.
This book represents the refereed proceedings of the Ninth International Conference on Monte Carlo and Quasi-Monte Carlo Methods in Scientific Computing that was held at the University of Warsaw (Poland) in August 2010. These biennial conferences are major events for Monte Carlo and the premiere event for quasi-Monte Carlo research. The proceedings include articles based on invited lectures as well as carefully selected contributed papers on all theoretical aspects and applications of Monte Carlo and quasi-Monte Carlo methods. The reader will be provided with information on latest developments in these very active areas. The book is an excellent reference for theoreticians and practitioners interested in solving high-dimensional computational problems arising, in particular, in finance and statistics.
The main goal of this book is to help organizations improve their effort estimates and effort estimation processes by providing a step-by-step methodology that takes them through the creation and validation of models that are based on their own knowledge and experience. Such models, once validated, can then be used to obtain predictions, carry out risk analyses, enhance their estimation processes for new projects and generally advance them as learning organizations. Emilia Mendes presents the Expert-Based Knowledge Engineering of Bayesian Networks (EKEBNs) methodology, which she has used and adapted during the course of several industry collaborations with different companies world-wide over more than 6 years. The book itself consists of two major parts: first, the methodology's foundations in knowledge management, effort estimation (with special emphasis on the intricacies of software and Web development) and Bayesian networks are detailed; then six industry case studies are presented which illustrate the practical use of EKEBNs. Domain experts from each company participated in the elicitation of the bespoke models for effort estimation and all models were built employing the widely-used Netica (TM) tool. This part is rounded off with a chapter summarizing the experiences with the methodology and the derived models. Practitioners working on software project management, software process quality or effort estimation and risk analysis in general will find a thorough introduction into an industry-proven methodology as well as numerous experiences, tips and possible pitfalls invaluable for their daily work.
People, problems, and proofs are the lifeblood of theoretical computer science. Behind the computing devices and applications that have transformed our lives are clever algorithms, and for every worthwhile algorithm there is a problem that it solves and a proof that it works. Before this proof there was an open problem: can one create an efficient algorithm to solve the computational problem? And, finally, behind these questions are the people who are excited about these fundamental issues in our computational world. In this book the authors draw on their outstanding research and teaching experience to showcase some key people and ideas in the domain of theoretical computer science, particularly in computational complexity and algorithms, and related mathematical topics. They show evidence of the considerable scholarship that supports this young field, and they balance an impressive breadth of topics with the depth necessary to reveal the power and the relevance of the work described. Beyond this, the authors discuss the sustained effort of their community, revealing much about the culture of their field. A career in theoretical computer science at the top level is a vocation: the work is hard, and in addition to the obvious requirements such as intellect and training, the vignettes in this book demonstrate the importance of human factors such as personality, instinct, creativity, ambition, tenacity, and luck. The authors' style is characterize d by personal observations, enthusiasm, and humor, and this book will be a source of inspiration and guidance for graduate students and researchers engaged with or planning careers in theoretical computer science.
This book presents selected papers from the 3rd International Workshop on Computational Engineering held in Stuttgart from October 6 to 10, 2014, bringing together innovative contributions from related fields with computer science and mathematics as an important technical basis among others. The workshop discussed the state of the art and the further evolution of numerical techniques for simulation in engineering and science. We focus on current trends in numerical simulation in science and engineering, new requirements arising from rapidly increasing parallelism in computer architectures, and novel mathematical approaches. Accordingly, the chapters of the book particularly focus on parallel algorithms and performance optimization, coupled systems, and complex applications and optimization.
The authors describe systematic methods for uncovering scientific laws a priori, on the basis of intuition, or "Gedanken Experiments". Mathematical expressions of scientific laws are, by convention, constrained by the rule that their form must be invariant with changes of the units of their variables. This constraint makes it possible to narrow down the possible forms of the laws. It is closely related to, but different from, dimensional analysis. It is a mathematical book, largely based on solving functional equations. In fact, one chapter is an introduction to the theory of functional equations.
Faithful communication is a necessary precondition for large-scale quantum information processing and networking, irrespective of the physical platform. Thus, the problems of quantum-state transfer and quantum-network engineering have attracted enormous interest over the last years, and constitute one of the most active areas of research in quantum information processing. The present volume introduces the reader to fundamental concepts and various aspects of this exciting research area, including links to other related areas and problems. The implementation of state-transfer schemes and the engineering of quantum networks are discussed in the framework of various quantum optical and condensed matter systems, emphasizing the interdisciplinary character of the research area. Each chapter is a review of theoretical or experimental achievements on a particular topic, written by leading scientists in the field. The volume aims at both newcomers as well as experienced researchers.
In this introductory textbook the author explains the key topics in cryptography. He takes a modern approach, where defining what is meant by "secure" is as important as creating something that achieves that goal, and security definitions are central to the discussion throughout. The author balances a largely non-rigorous style - many proofs are sketched only - with appropriate formality and depth. For example, he uses the terminology of groups and finite fields so that the reader can understand both the latest academic research and "real-world" documents such as application programming interface descriptions and cryptographic standards. The text employs colour to distinguish between public and private information, and all chapters include summaries and suggestions for further reading. This is a suitable textbook for advanced undergraduate and graduate students in computer science, mathematics and engineering, and for self-study by professionals in information security. While the appendix summarizes most of the basic algebra and notation required, it is assumed that the reader has a basic knowledge of discrete mathematics, probability, and elementary calculus.
This revised and expanded new edition elucidates the elegance and simplicity of the fundamental theory underlying formal languages and compilation. Retaining the reader-friendly style of the 1st edition, this versatile textbook describes the essential principles and methods used for defining the syntax of artificial languages, and for designing efficient parsing algorithms and syntax-directed translators with semantic attributes. Features: presents a novel conceptual approach to parsing algorithms that applies to extended BNF grammars, together with a parallel parsing algorithm (NEW); supplies supplementary teaching tools at an associated website; systematically discusses ambiguous forms, allowing readers to avoid pitfalls; describes all algorithms in pseudocode; makes extensive usage of theoretical models of automata, transducers and formal grammars; includes concise coverage of algorithms for processing regular expressions and finite automata; introduces static program analysis based on flow equations.
Meshfree methods, particle methods, and generalized finite element methods have witnessed substantial development since the mid 1990s. The growing interest in these methods is due in part to the fact that they are extremely flexible numerical tools and can be interpreted in a number of ways. For instance, meshfree methods can be viewed as a natural extension of classical finite element and finite difference methods to scattered node configurations with no fixed connectivity. Furthermore, meshfree methods offer a number of advantageous features which are especially attractive when dealing with multiscale phenomena: a priori knowledge about particular local behavior of the solution can easily be introduced in the meshfree approximation space, and coarse-scale approximations can be seamlessly refined with fine-scale information. This volume collects selected papers presented at the Seventh International Workshop on Meshfree Methods, held in Bonn, Germany in September 2013. They address various aspects of this highly dynamic research field and cover topics from applied mathematics, physics and engineering.
This unique textbook/reference presents unified coverage of bioinformatics topics relating to both biological sequences and biological networks, providing an in-depth analysis of cutting-edge distributed algorithms, as well as of relevant sequential algorithms. In addition to introducing the latest algorithms in this area, more than fifteen new distributed algorithms are also proposed. Topics and features: reviews a range of open challenges in biological sequences and networks; describes in detail both sequential and parallel/distributed algorithms for each problem; suggests approaches for distributed algorithms as possible extensions to sequential algorithms, when the distributed algorithms for the topic are scarce; proposes a number of new distributed algorithms in each chapter, to serve as potential starting points for further research; concludes each chapter with self-test exercises, a summary of the key points, a comparison of the algorithms described, and a literature review. |
![]() ![]() You may like...
Activate the Third Space - How to Align…
Bill Cornwell, Michael Switow
Hardcover
Practical Guide to Sperm Analysis…
Nicolas Garrido, Rocio Rivera
Hardcover
R5,372
Discovery Miles 53 720
|