![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > General theory of computing > Mathematical theory of computation
This book is primarily intended as a research monograph that could also be used in graduate courses for the design of parallel algorithms in matrix computations. It assumes general but not extensive knowledge of numerical linear algebra, parallel architectures, and parallel programming paradigms. The book consists of four parts: (I) Basics; (II) Dense and Special Matrix Computations; (III) Sparse Matrix Computations; and (IV) Matrix functions and characteristics. Part I deals with parallel programming paradigms and fundamental kernels, including reordering schemes for sparse matrices. Part II is devoted to dense matrix computations such as parallel algorithms for solving linear systems, linear least squares, the symmetric algebraic eigenvalue problem, and the singular-value decomposition. It also deals with the development of parallel algorithms for special linear systems such as banded ,Vandermonde ,Toeplitz ,and block Toeplitz systems. Part III addresses sparse matrix computations: (a) the development of parallel iterative linear system solvers with emphasis on scalable preconditioners, (b) parallel schemes for obtaining a few of the extreme eigenpairs or those contained in a given interval in the spectrum of a standard or generalized symmetric eigenvalue problem, and (c) parallel methods for computing a few of the extreme singular triplets. Part IV focuses on the development of parallel algorithms for matrix functions and special characteristics such as the matrix pseudospectrum and the determinant. The book also reviews the theoretical and practical background necessary when designing these algorithms and includes an extensive bibliography that will be useful to researchers and students alike. The book brings together many existing algorithms for the fundamental matrix computations that have a proven track record of efficient implementation in terms of data locality and data transfer on state-of-the-art systems, as well as several algorithms that are presented for the first time, focusing on the opportunities for parallelism and algorithm robustness.
This book starts by introducing the fundamental concepts of mathematical continuum mechanics for fluids and solids and their coupling. Special attention is given to the derivation of variational formulations for the subproblems describing fluid- and solid-mechanics as well as the coupled fluid-structure interaction problem. Two monolithic formulations for fluid-structure interactions are described in detail: the well-established ALE formulation and the modern Fully Eulerian formulation, which can effectively deal with problems featuring large deformation and contact. Further, the book provides details on state-of-the-art discretization schemes for fluid- and solid-mechanics and considers the special needs of coupled problems with interface-tracking and interface-capturing techniques. Lastly, advanced topics like goal-oriented error estimation, multigrid solution and gradient-based optimization schemes are discussed in the context of fluid-structure interaction problems.
This research monograph focuses on the design of arithmetic circuits in Quantum Dot Cellular Automata (QCA). Using the fact that the 3-input majority gate is a primitive in QCA, the book sets out to discover hitherto unknown properties of majority logic in the context of arithmetic circuit designs. The pursuit for efficient adders in QCA takes two forms. One involves application of the new results in majority logic to existing adders. The second involves development of a custom adder for QCA technology. A QCA adder named as hybrid adder is proposed and it is shown that it outperforms existing multi-bit adders with respect to area and delay. The work is extended to the design of a low-complexity multiplier for signed numbers in QCA. Furthermore the book explores two aspects unique to QCA technology, namely thermal robustness and the role of interconnects. In addition, the book introduces the reader to QCA layout design and simulation using QCADesigner. Features & Benefits: This research-based book: *Introduces the reader to Quantum Dot Cellular Automata, an emerging nanotechnology. *Explores properties of majority logic. *Demonstrates application of the properties to design efficient arithmetic circuits. *Guides the reader towards layout design and simulation in QCADesigner.
This book presents the latest developments regarding a detailed mobile agent-enabled anomaly detection and verification system for resource constrained sensor networks; a number of algorithms on multi-aspect anomaly detection in sensor networks; several algorithms on mobile agent transmission optimization in resource constrained sensor networks; an algorithm on mobile agent-enabled in situ verification of anomalous sensor nodes; a detailed Petri Net-based formal modeling and analysis of the proposed system, and an algorithm on fuzzy logic-based cross-layer anomaly detection and mobile agent transmission optimization. As such, it offers a comprehensive text for interested readers from academia and industry alike.
This book reports on the latest advances and applications of chaotic systems. It consists of 25 contributed chapters by experts who are specialized in the various topics addressed in this book. The chapters cover a broad range of topics of chaotic systems such as chaos, hyperchaos, jerk systems, hyperjerk systems, conservative and dissipative systems, circulant chaotic systems, multi-scroll chaotic systems, finance chaotic system, highly chaotic systems, chaos control, chaos synchronization, circuit realization and applications of chaos theory in secure communications, mobile robot, memristors, cellular neural networks, etc. Special importance was given to chapters offering practical solutions, modeling and novel control methods for the recent research problems in chaos theory. This book will serve as a reference book for graduate students and researchers with a basic knowledge of chaos theory and control systems. The resulting design procedures on the chaotic systems are emphasized using MATLAB software.
This book is aimed at presenting concepts, methods and algorithms ableto cope with undersampled and limited data. One such trend that recently gained popularity and to some extent revolutionised signal processing is compressed sensing. Compressed sensing builds upon the observation that many signals in nature are nearly sparse (or compressible, as they are normally referred to) in some domain, and consequently they can be reconstructed to within high accuracy from far fewer observations than traditionally held to be necessary. Apart from compressed sensing this book contains other related approaches. Each methodology has its own formalities for dealing with such problems. As an example, in the Bayesian approach, sparseness promoting priors such as Laplace and Cauchy are normally used for penalising improbable model variables, thus promoting low complexity solutions. Compressed sensing techniques and homotopy-type solutions, such as the LASSO, utilise l1-norm penalties for obtaining sparse solutions using fewer observations than conventionally needed. The book emphasizes on the role of sparsity as a machinery for promoting low complexity representations and likewise its connections to variable selection and dimensionality reduction in various engineering problems. This book is intended for researchers, academics and practitioners with interest in various aspects and applications of sparse signal processing.
This book shows cognitive scientists in training how mathematics, computer science and science can be usefully and seamlessly intertwined. It is a follow-up to the first two volumes on mathematics for cognitive scientists, and includes the mathematics and computational tools needed to understand how to compute the terms in the Fourier series expansions that solve the cable equation. The latter is derived from first principles by going back to cellular biology and the relevant biophysics. A detailed discussion of ion movement through cellular membranes, and an explanation of how the equations that govern such ion movement leading to the standard transient cable equation are included. There are also solutions for the cable model using separation of variables, as well an explanation of why Fourier series converge and a description of the implementation of MatLab tools to compute the solutions. Finally, the standard Hodgkin - Huxley model is developed for an excitable neuron and is solved using MatLab.
This book gathers threads that have evolved across different mathematical disciplines into seamless narrative. It deals with condition as a main aspect in the understanding of the performance ---regarding both stability and complexity--- of numerical algorithms. While the role of condition was shaped in the last half-century, so far there has not been a monograph treating this subject in a uniform and systematic way. The book puts special emphasis on the probabilistic analysis of numerical algorithms via the analysis of the corresponding condition. The exposition's level increases along the book, starting in the context of linear algebra at an undergraduate level and reaching in its third part the recent developments and partial solutions for Smale's 17th problem which can be explained within a graduate course. Its middle part contains a condition-based course on linear programming that fills a gap between the current elementary expositions of the subject based on the simplex method and those focusing on convex programming.
This book is intended to make recent results on the derivation of higher order numerical schemes for random ordinary differential equations (RODEs) available to a broader readership, and to familiarize readers with RODEs themselves as well as the closely associated theory of random dynamical systems. In addition, it demonstrates how RODEs are being used in the biological sciences, where non-Gaussian and bounded noise are often more realistic than the Gaussian white noise in stochastic differential equations (SODEs). RODEs are used in many important applications and play a fundamental role in the theory of random dynamical systems. They can be analyzed pathwise with deterministic calculus, but require further treatment beyond that of classical ODE theory due to the lack of smoothness in their time variable. Although classical numerical schemes for ODEs can be used pathwise for RODEs, they rarely attain their traditional order since the solutions of RODEs do not have sufficient smoothness to have Taylor expansions in the usual sense. However, Taylor-like expansions can be derived for RODEs using an iterated application of the appropriate chain rule in integral form, and represent the starting point for the systematic derivation of consistent higher order numerical schemes for RODEs. The book is directed at a wide range of readers in applied and computational mathematics and related areas as well as readers who are interested in the applications of mathematical models involving random effects, in particular in the biological sciences.The level of this book is suitable for graduate students in applied mathematics and related areas, computational sciences and systems biology. A basic knowledge of ordinary differential equations and numerical analysis is required.
This book questions the relevance of computation to the physical universe. Our theories deliver computational descriptions, but the gaps and discontinuities in our grasp suggest a need for continued discourse between researchers from different disciplines, and this book is unique in its focus on the mathematical theory of incomputability and its relevance for the real world. The core of the book consists of thirteen chapters in five parts on extended models of computation; the search for natural examples of incomputable objects; mind, matter, and computation; the nature of information, complexity, and randomness; and the mathematics of emergence and morphogenesis. This book will be of interest to researchers in the areas of theoretical computer science, mathematical logic, and philosophy.
In this monograph we introduce and examine four new temporal logic formalisms that can be used as specification languages for the automated verification of the reliability of hardware and software designs with respect to a desired behavior. The work is organized in two parts. In the first part two logics for computations, the graded computation tree logic and the computation tree logic with minimal model quantifiers are discussed. These have proved to be useful in describing correct executions of monolithic closed systems. The second part focuses on logics for strategies, strategy logic and memoryful alternating-time temporal logic, which have been successfully applied to formalize several properties of interactive plays in multi-entities systems modeled as multi-agent games.
This book illustrates how to use description logic-based formalisms to their full potential in the creation, indexing, and reuse of multimedia semantics. To do so, it introduces researchers to multimedia semantics by providing an in-depth review of state-of-the-art standards, technologies, ontologies, and software tools. It draws attention to the importance of formal grounding in the knowledge representation of multimedia objects, the potential of multimedia reasoning in intelligent multimedia applications, and presents both theoretical discussions and best practices in multimedia ontology engineering. Readers already familiar with mathematical logic, Internet, and multimedia fundamentals will learn to develop formally grounded multimedia ontologies, and map concept definitions to high-level descriptors. The core reasoning tasks, reasoning algorithms, and industry-leading reasoners are presented, while scene interpretation via reasoning is also demonstrated. Overall, this book offers readers an essential introduction to the formal grounding of web ontologies, as well as a comprehensive collection and review of description logics (DLs) from the perspectives of expressivity and reasoning complexity. It covers best practices for developing multimedia ontologies with formal grounding to guarantee decidability and obtain the desired level of expressivity while maximizing the reasoning potential. The capabilities of such multimedia ontologies are demonstrated by DL implementations with an emphasis on multimedia reasoning applications.
This book opens the door to a new interesting and ambitious world of reversible and quantum computing research. It presents the state of the art required to travel around that world safely. Top world universities, companies and government institutions are in a race of developing new methodologies, algorithms and circuits on reversible logic, quantum logic, reversible and quantum computing and nano-technologies. In this book, twelve reversible logic synthesis methodologies are presented for the first time in a single literature with some new proposals. Also, the sequential reversible logic circuitries are discussed for the first time in a book. Reversible logic plays an important role in quantum computing. Any progress in the domain of reversible logic can be directly applied to quantum logic. One of the goals of this book is to show the application of reversible logic in quantum computing. A new implementation of wavelet and multiwavelet transforms using quantum computing is performed for this purpose. Researchers in academia or industry and graduate students, who work in logic synthesis, quantum computing, nano-technology, and low power VLSI circuit design, will be interested in this book.
The papers in this volume represent the most timely and advanced contributions to the 2014 Joint Applied Statistics Symposium of the International Chinese Statistical Association (ICSA) and the Korean International Statistical Society (KISS), held in Portland, Oregon. The contributions cover new developments in statistical modeling and clinical research: including model development, model checking, and innovative clinical trial design and analysis. Each paper was peer-reviewed by at least two referees and also by an editor. The conference was attended by over 400 participants from academia, industry, and government agencies around the world, including from North America, Asia, and Europe. It offered 3 keynote speeches, 7 short courses, 76 parallel scientific sessions, student paper sessions, and social events.
This book offers a self-study program on how mathematics, computer science and science can be profitably and seamlessly intertwined. This book focuses on two variable ODE models, both linear and nonlinear, and highlights theoretical and computational tools using MATLAB to explain their solutions. It also shows how to solve cable models using separation of variables and the Fourier Series.
This monograph addresses the state of the art of reduced order methods for modeling and computational reduction of complex parametrized systems, governed by ordinary and/or partial differential equations, with a special emphasis on real time computing techniques and applications in computational mechanics, bioengineering and computer graphics. Several topics are covered, including: design, optimization, and control theory in real-time with applications in engineering; data assimilation, geometry registration, and parameter estimation with special attention to real-time computing in biomedical engineering and computational physics; real-time visualization of physics-based simulations in computer science; the treatment of high-dimensional problems in state space, physical space, or parameter space; the interactions between different model reduction and dimensionality reduction approaches; the development of general error estimation frameworks which take into account both model and discretization effects. This book is primarily addressed to computational scientists interested in computational reduction techniques for large scale differential problems.
This volume presents the latest advances and trends in stochastic models and related statistical procedures. Selected peer-reviewed contributions focus on statistical inference, quality control, change-point analysis and detection, empirical processes, time series analysis, survival analysis and reliability, statistics for stochastic processes, big data in technology and the sciences, statistical genetics, experiment design, and stochastic models in engineering. Stochastic models and related statistical procedures play an important part in furthering our understanding of the challenging problems currently arising in areas of application such as the natural sciences, information technology, engineering, image analysis, genetics, energy and finance, to name but a few. This collection arises from the 12th Workshop on Stochastic Models, Statistics and Their Applications, Wroclaw, Poland.
This book provides a critical examination of how the choice of what to believe is represented in the standard model of belief change. In particular the use of possible worlds and infinite remainders as objects of choice is critically examined. Descriptors are introduced as a versatile tool for expressing the success conditions of belief change, addressing both local and global descriptor revision. The book presents dynamic descriptors such as Ramsey descriptors that convey how an agent's beliefs tend to be changed in response to different inputs. It also explores sentential revision and demonstrates how local and global operations of revision by a sentence can be derived as a special case of descriptor revision. Lastly, the book examines revocation, a generalization of contraction in which a specified sentence is removed in a process that may possibly also involve the addition of some new information to the belief set.
This book features selected papers presented at the 2nd International Conference on Advanced Computing Technologies and Applications, held at SVKM's Dwarkadas J. Sanghvi College of Engineering, Mumbai, India, from 28 to 29 February 2020. Covering recent advances in next-generation computing, the book focuses on recent developments in intelligent computing, such as linguistic computing, statistical computing, data computing and ambient applications.
This thesis deals with topological orders from two different perspectives: from a condensed matter point of view, where topological orders are considered as breakthrough phases of matter; and from the emerging realm of quantum computation, where topological quantum codes are considered the most appealing platform against decoherence. The thesis reports remarkable studies from both sides. It thoroughly investigates a topological order called the double semion model, a counterpart of the Kitaev model but exhibiting richer quasiparticles as excitations. A new model for symmetry enriched topological order is constructed, which adds an onsite global symmetry to the double semion model. Using this topological phase, a new example of topological code is developed, the semion code, which is non-CSS, additive, non-Pauli and within the stabiliser formalism. Furthermore, the thesis analyses the Rashba spin-orbit coupling within topological insulators, turning the helical edge states into generic edges modes with potential application in spinstronics. New types of topological superconductors are proposed and the novel properties of the correspondingly created Majorana fermions are investigated. These Majorana fermions have inherent properties enabling braiding and the performance of logical gates as fundamental blocks for a universsal quantum computator.
This book offers an in-depth insight into the general-purpose finite element program MSC Marc, which is distributed by MSC Software Corporation. It is a specialized program for nonlinear problems (implicit solver) which is common in academia and industry. The primary goal of this book is to provide a comprehensive introduction to a special feature of this software: the user can write user-subroutines in the programming language Fortran, which is the language of all classical finite element packages. This subroutine feature allows the user to replace certain modules of the core code and to implement new features such as constitutive laws or new elements. Thus, the functionality of commercial codes ('black box') can easily be extended by linking user written code to the main core of the program. This feature allows to take advantage of a commercial software package with the flexibility of a 'semi-open' code.
In a fragment entitled Elementa Nova Matheseos Universalis (1683?) Leibniz writes "the mathesis [...] shall deliver the method through which things that are conceivable can be exactly determined"; in another fragment he takes the mathesis to be "the science of all things that are conceivable." Leibniz considers all mathematical disciplines as branches of the mathesis and conceives the mathesis as a general science of forms applicable not only to magnitudes but to every object that exists in our imagination, i.e. that is possible at least in principle. As a general science of forms the mathesis investigates possible relations between "arbitrary objects" ("objets quelconques"). It is an abstract theory of combinations and relations among objects whatsoever. In 1810 the mathematician and philosopher Bernard Bolzano published a booklet entitled Contributions to a Better-Grounded Presentation of Mathematics. There is, according to him, a certain objective connection among the truths that are germane to a certain homogeneous field of objects: some truths are the "reasons" ("Grunde") of others, and the latter are "consequences" ("Folgen") of the former. The reason-consequence relation seems to be the counterpart of causality at the level of a relation between true propositions. Arigorous proof is characterized in this context as a proof that shows the reason of the proposition that is to be proven. Requirements imposed on rigorous proofs seem to anticipate normalization results in current proof theory. The contributors of Mathesis Universalis, Computability and Proof, leading experts in the fields of computer science, mathematics, logic and philosophy, show the evolution of these and related ideas exploring topics in proof theory, computability theory, intuitionistic logic, constructivism and reverse mathematics, delving deeply into a contextual examination of the relationship between mathematical rigor and demands for simplification.
This book primarily addresses Intelligent Information Systems (IIS) and the integration of artificial intelligence, intelligent systems and technologies, database technologies and information systems methodologies to create the next generation of information systems. It includes original and state-of-the-art research on theoretical and practical advances in IIS, system architectures, tools and techniques, as well as "success stories" in intelligent information systems. Intended as an interdisciplinary forum in which scientists and professionals could share their research results and report on new developments and advances in intelligent information systems, technologies and related areas - as well as their applications - , it offers a valuable resource for researchers and practitioners alike.
This book offers an introduction to applications prompted by tensor analysis, especially by the spectral tensor theory developed in recent years. It covers applications of tensor eigenvalues in multilinear systems, exponential data fitting, tensor complementarity problems, and tensor eigenvalue complementarity problems. It also addresses higher-order diffusion tensor imaging, third-order symmetric and traceless tensors in liquid crystals, piezoelectric tensors, strong ellipticity for elasticity tensors, and higher-order tensors in quantum physics. This book is a valuable reference resource for researchers and graduate students who are interested in applications of tensor eigenvalues.
The papers in this volume represent a broad, applied swath of advanced contributions to the 2015 ICSA/Graybill Applied Statistics Symposium of the International Chinese Statistical Association, held at Colorado State University in Fort Collins. The contributions cover topics that range from statistical applications in business and finance to applications in clinical trials and biomarker analysis. Each papers was peer-reviewed by at least two referees and also by an editor. The conference was attended by over 400 participants from academia, industry, and government agencies around the world, including from North America, Asia, and Europe. |
![]() ![]() You may like...
Predictive Analytics in Cloud, Fog, and…
Hiren Kumar Thakkar, Chinmaya Kumar Dehury, …
Hardcover
R4,927
Discovery Miles 49 270
Computational Intelligence for Big Data…
D P Acharjya, Satchidananda Dehuri, …
Hardcover
All About Music - The Complete Ontology…
Guerino Mazzola, Maria Mannone, …
Hardcover
R2,278
Discovery Miles 22 780
On Fuzziness - A Homage to Lotfi A…
Rudolf Seising, Enric Trillas, …
Hardcover
R5,179
Discovery Miles 51 790
Creating and Digitizing Language Corpora…
Karen P. Corrigan, Adam Mearns
Hardcover
|