![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > General theory of computing > Mathematical theory of computation
This book presents the theory of continuum mechanics for mechanical, thermodynamical, and electrodynamical systems. It shows how to obtain governing equations and it applies them by computing the reality. It uses only open-source codes developed under the FEniCS project and includes codes for 20 engineering applications from mechanics, fluid dynamics, applied thermodynamics, and electromagnetism. Moreover, it derives and utilizes the constitutive equations including coupling terms, which allow to compute multiphysics problems by incorporating interactions between primitive variables, namely, motion, temperature, and electromagnetic fields. An engineering system is described by the primitive variables satisfying field equations that are partial differential equations in space and time. The field equations are mostly coupled and nonlinear, in other words, difficult to solve. In order to solve the coupled, nonlinear system of partial differential equations, the book uses a novel collection of open-source packages developed under the FEniCS project. All primitive variables are solved at once in a fully coupled fashion by using finite difference method in time and finite element method in space.
This bookisan outgrowthoften yearsof researchatthe Universityof Florida Computational NeuroEngineering Laboratory (CNEL) in the general area of statistical signal processing and machine learning. One of the goals of writing the book is exactly to bridge the two ?elds that share so many common problems and techniques but are not yet e?ectively collaborating. Unlikeotherbooks thatcoverthe state ofthe artinagiven?eld, this book cuts across engineering (signal processing) and statistics (machine learning) withacommontheme: learningseenfromthepointofviewofinformationt- orywithanemphasisonRenyi'sde?nitionofinformation.Thebasicapproach is to utilize the information theory descriptors of entropy and divergence as nonparametric cost functions for the design of adaptive systems in unsup- vised or supervised training modes. Hence the title: Information-Theoretic Learning (ITL). In the course of these studies, we discovered that the main idea enabling a synergistic view as well as algorithmic implementations, does not involve the conventional central moments of the data (mean and covariance). Rather, the core concept is the ?-norm of the PDF, in part- ular its expected value (? = 2), which we call the information potential. This operator and related nonparametric estimators link information theory, optimization of adaptive systems, and reproducing kernel Hilbert spaces in a simple and unconventional way.
Digital arithmetic plays an important role in the design of
general-purpose digital processors and of embedded systems for
signal processing, graphics, and communications. In spite of a
mature body of knowledge in digital arithmetic, each new generation
of processors or digital systems creates new arithmetic design
problems. Designers, researchers, and graduate students will find
solid solutions to these problems in this comprehensive,
state-of-the-art exposition of digital arithmetic.
This volume contains thirteen articles on advances in applied mathematics and computing methods for engineering problems. Six papers are on optimization methods and algorithms with emphasis on problems with multiple criteria; four articles are on numerical methods for applied problems modeled with nonlinear PDEs; two contributions are on abstract estimates for error analysis; finally one paper deals with rare events in the context of uncertainty quantification. Applications include aerospace, glaciology and nonlinear elasticity. Hereinis a selection of contributions from speakers at two conferences on applied mathematics held in June 2012 at the University of Jyvaskyla, Finland. The first conference, Optimization and PDEs with Industrial Applications celebrated the seventieth birthday of Professor Jacques Periaux of theUniversity of Jyvaskyla and Polytechnic University of Catalonia (Barcelona Tech) and the second conference, Optimization and PDEs with Applications celebrated the seventy-fifth birthday of Professor Roland Glowinski of the University of Houston. This work should be of interest to researchers and practitioners as well as advanced students or engineers in computational and applied mathematics or mechanics."
This volume gathers papers presented at the Workshop on Computational Diffusion MRI (CDMRI'18), which was held under the auspices of the International Conference on Medical Image Computing and Computer Assisted Intervention in Granada, Spain on September 20, 2018. It presents the latest developments in the highly active and rapidly growing field of diffusion MRI. The reader will find papers on a broad range of topics, from the mathematical foundations of the diffusion process and signal generation, to new computational methods and estimation techniques for the in-vivo recovery of microstructural and connectivity features, as well as harmonisation and frontline applications in research and clinical practice. The respective papers constitute invited works from high-profile researchers with a specific focus on three topics that are now gaining momentum within the diffusion MRI community: i) machine learning for diffusion MRI; ii) diffusion MRI outside the brain (e.g. in the placenta); and iii) diffusion MRI for multimodal imaging. The book shares new perspectives on the latest research challenges for those currently working in the field, but also offers a valuable starting point for anyone interested in learning computational techniques in diffusion MRI. It includes rigorous mathematical derivations, a wealth of full-colour visualisations, and clinically relevant results. As such, it will be of interest to researchers and practitioners in the fields of computer science, MRI physics and applied mathematics alike.
This book presents task-scheduling techniques for emerging complex parallel architectures including heterogeneous multi-core architectures, warehouse-scale datacenters, and distributed big data processing systems. The demand for high computational capacity has led to the growing popularity of multicore processors, which have become the mainstream in both the research and real-world settings. Yet to date, there is no book exploring the current task-scheduling techniques for the emerging complex parallel architectures. Addressing this gap, the book discusses state-of-the-art task-scheduling techniques that are optimized for different architectures, and which can be directly applied in real parallel systems. Further, the book provides an overview of the latest advances in task-scheduling policies in parallel architectures, and will help readers understand and overcome current and emerging issues in this field.
This book discusses recent developments and contemporary research in mathematics, statistics and their applications in computing. All contributing authors are eminent academicians, scientists, researchers and scholars in their respective fields, hailing from around the world. The conference has emerged as a powerful forum, offering researchers a venue to discuss, interact and collaborate and stimulating the advancement of mathematics and its applications in computer science. The book will allow aspiring researchers to update their knowledge of cryptography, algebra, frame theory, optimizations, stochastic processes, compressive sensing, functional analysis, complex variables, etc. Educating future consumers, users, producers, developers and researchers in mathematics and computing is a challenging task and essential to the development of modern society. Hence, mathematics and its applications in computer science are of vital importance to a broad range of communities, including mathematicians and computing professionals across different educational levels and disciplines.
Refinement is one of the cornerstones of the formal approach to software engineering, and its use in various domains has led to research on new applications and generalisation. This book brings together this important research in one volume, with the addition of examples drawn from different application areas. It covers four main themes: Data refinement and its application to Z Generalisations of refinement that change the interface and atomicity of operations Refinement in Object-Z Modelling state and behaviour by combining Object-Z with CSP Refinement in Z and Object-Z: Foundations and Advanced Applications provides an invaluable overview of recent research for academic and industrial researchers, lecturers teaching formal specification and development, industrial practitioners using formal methods in their work, and postgraduate and advanced undergraduate students. This second edition is a comprehensive update to the first and includes the following new material: Early chapters have been extended to also include trace refinement, based directly on partial relations rather than through totalisation Provides an updated discussion on divergence, non-atomic refinements and approximate refinement Includes a discussion of the differing semantics of operations and outputs and how they affect the abstraction of models written using Object-Z and CSP Presents a fuller account of the relationship between relational refinement and various models of refinement in CSP Bibliographic notes at the end of each chapter have been extended with the most up to date citations and research
"Set Theory for Computing" provides a comprehensive account of set-oriented symbolic manipulation methods suitable for automated reasoning. Its main objective is twofold: 1) to provide a flexible formalization for a variety of set languages, and 2) to clarify the semantics of set constructs firmly established in modern specification languages and in the programming practice. Topics include: semantic unification, decision algorithms, modal logics, declarative programming, tableau-based proof techniques, and theory-based theorem proving. The style of presentation is self-contained, rigorous and accurate. Some familiarity with symbolic logic is helpful but not a requirement. This book is a useful resource for all advanced students, professionals, and researchers in computing sciences, artificial intelligence, automated reasoning, logic, and computational mathematics. It will serve to complement their intuitive understanding of set concepts with the ability to master them by symbolic and logically based algorithmic methods and deductive techniques.
This book describes a novel methodology for studying algorithmic skills, intended as cognitive activities related to rule-based symbolic transformation, and argues that some human computational abilities may be interpreted and analyzed as genuine examples of extended cognition. It shows that the performance of these abilities relies not only on innate neurocognitive systems or language-related skills, but also on external tools and general agent-environment interactions. Further, it asserts that a low-level analysis, based on a set of core neurocognitive systems linking numbers and language, is not sufficient to explain some specific forms of high-level numerical skills, like those involved in algorithm execution. To this end, it reports on the design of a cognitive architecture for modeling all the relevant features involved in the execution of algorithmic strategies, including external tools, such as paper and pencils. The first part of the book discusses the philosophical premises for endorsing and justifying a position in philosophy of mind that links a modified form of computationalism with some recent theoretical and scientific developments, like those introduced by the so-called dynamical approach to cognition. The second part is dedicated to the description of a Turing-machine-inspired cognitive architecture, expressly designed to formalize all kinds of algorithmic strategies.
This volume, the 6th volume in the DRUMS Handbook series, is part of the after math of the successful ESPRIT project DRUMS (Defeasible Reasoning and Un certainty Management Systems) which took place in two stages from 1989-1996. In the second stage (1993-1996) a work package was introduced devoted to the topics Reasoning and Dynamics, covering both the topics of 'Dynamics of Rea soning', where reasoning is viewed as a process, and 'Reasoning about Dynamics', which must be understood as pertaining to how both designers of and agents within dynamic systems may reason about these systems. The present volume presents work done in this context. This work has an emphasis on modelling and formal techniques in the investigation of the topic "Reasoning and Dynamics," but it is not mere theory that occupied us. Rather research was aimed at bridging the gap between theory and practice. Therefore also real-life applications of the modelling techniques were considered, and we hope this also shows in this volume, which is focused on the dynamics of reasoning processes. In order to give the book a broader perspective, we have invited a number of well-known researchers outside the project but working on similar topics to contribute as well. We have very pleasant recollections of the project, with its lively workshops and other meetings, with the many sites and researchers involved, both within and outside our own work package."
Integral equations have wide applications in various fields, including continuum mechanics, potential theory, geophysics, electricity and magnetism, kinetic theory of gases, hereditary phenomena in physics and biology, renewal theory, quantum mechanics, radiation, optimization, optimal control systems, communication theory, mathematical economics, population genetics, queueing theory, and medicine. Computational Methods for Linear Integral Equations presents basic theoretical material that deals with numerical analysis, convergence, error estimates, and accuracy. The unique computational aspect leads the reader from theoretical and practical problems all the way through to computation with hands-on guidance for input files and the execution of computer programs. Features: * Offers all supporting MathematicaA(R) files related to the book via the Internet at the authors' Web sites: www.math.uno.edu/fac/pkythe.html or www.math.uno.edu/fac/ppuri.html * Contains identification codes for problems, related methods, and computer programs that are cross-referenced throughout the book to make the connections easy to understand * Illustrates a how-to approach to computational work in the development of algorithms, construction of input files, timing, and accuracy analysis * Covers linear integral equations of Fredholm and Volterra types of the first and second kinds as well as associated singular integral equations, integro-differential equations, and eigenvalue problems * Provides clear, step-by-step guidelines for solving difficult and complex computational problems This book is an essential reference and authoritative resource for all professionals, graduate students, and researchers in mathematics, physical sciences, and engineering. Researchers interested in the numerical solution of integral equations will find its practical problem-solving style both accessible and useful for their work.
This book is a short, concise introduction to key mathematical ideas for computing students which develops their understanding of discrete mathematics and its application in computing. The topics are presented in a well defined, logical order that build upon each other and are constantly reinforced by worked examples. Reliance on students' previous mathematical experience is kept to a minimum, though some basic algebraic manipulation is required. This book is appropriate for CS and Math students in an undergraduate Discrete Math course. The content constitutes an accepted core of mathematics for computer scientists (for example, the formal methods used in computer science draw heavily on the discrete methematical concepts covered here, particularly logic, sets, relations and functions). Emphasis is placed on clear and careful explanations of basic ideas and on building confidence in developing mathematical competence through carefully selected exercises. All chapters conclude with short applications/case studies relevant to computing, which provide further motivation to engage with the mathematical ideas involved, and also demonstrate how the mathematics can be applied in a computing context.
The Mathieu series is a functional series introduced by Emile Leonard Mathieu for the purposes of his research on the elasticity of solid bodies. Bounds for this series are needed for solving biharmonic equations in a rectangular domain. In addition to Tomovski and his coauthors, Pogany, Cerone, H. M. Srivastava, J. Choi, etc. are some of the known authors who published results concerning the Mathieu series, its generalizations and their alternating variants. Applications of these results are given in classical, harmonic and numerical analysis, analytical number theory, special functions, mathematical physics, probability, quantum field theory, quantum physics, etc. Integral representations, analytical inequalities, asymptotic expansions and behaviors of some classes of Mathieu series are presented in this book. A systematic study of probability density functions and probability distributions associated with the Mathieu series, its generalizations and Planck's distribution is also presented. The book is addressed at graduate and PhD students and researchers in mathematics and physics who are interested in special functions, inequalities and probability distributions.
The topic of level sets is currently very timely and useful for creating realistic 3-D images and animations. They are powerful numerical techniques for analyzing and computing interface motion in a host of application settings. In computer vision, it has been applied to stereo and segmentation, whereas in graphics it has been applied to the postproduction process of in-painting and 3-D model construction. Osher is co-inventor of the Level Set Methods, a pioneering framework introduced jointly with James Sethian from the University of Berkeley in 1998. This methodology has been used up to now to provide solutions to a wide application range not limited to image processing, computer vision, robotics, fluid mechanics, crystallography, lithography, and computer graphics. The topic is of great interest to advanced students, professors, and R&D professionals working in the areas of graphics (post-production), video-based surveillance, visual inspection, augmented reality, document image processing, and medical image processing. These techniques are already employed to provide solutions and products in the industry (Cognitech, Siemens, Philips, Focus Imaging). An essential compilation of survey chapters from the leading researchers in the field, emphasizing the applications of the methods. This book can be suitable for a short professional course related with the processing of visual information.
This book presents a deep spectrum of musical, mathematical, physical, and philosophical perspectives that have emerged in this field at the intersection of music and mathematics. In particular the contributed chapters introduce advanced techniques and concepts from modern mathematics and physics, deriving from successes in domains such as Topos theory and physical string theory. The authors include many of the leading researchers in this domain, and the book will be of value to researchers working in computational music, particularly in the areas of counterpoint, gesture, and Topos theory.
This book on proof theory centers around the legacy of Kurt Schutte and its current impact on the subject. Schutte was the last doctoral student of David Hilbert who was the first to see that proofs can be viewed as structured mathematical objects amenable to investigation by mathematical methods (metamathematics). Schutte inaugurated the important paradigm shift from finite proofs to infinite proofs and developed the mathematical tools for their analysis. Infinitary proof theory flourished in his hands in the 1960s, culminating in the famous bound 0 for the limit of predicative mathematics (a fame shared with Feferman). Later his interests shifted to developing infinite proof calculi for impredicative theories. Schutte had a keen interest in advancing ordinal analysis to ever stronger theories and was still working on some of the strongest systems in his eighties. The articles in this volume from leading experts close to his research, show the enduring influence of his work in modern proof theory. They range from eye witness accounts of his scientific life to developments at the current research frontier, including papers by Schutte himself that have never been published before.
All over the world sport plays a prominent role in society: as a leisure activity for many, as an ingredient of culture, as a business and as a matter of national prestige in such major events as the World Cup in soccer or the Olympic Games. Hence, it is not surprising that science has entered the realm of sports, and, in particular, that computer simulation has become highly relevant in recent years. This is explored in this book by choosing five different sports as examples, demonstrating that computational science and engineering (CSE) can make essential contributions to research on sports topics on both the fundamental level and, eventually, by supporting athletes performance."
The book provides a bottom-up approach to understanding how a computer works and how to use computing to solve real-world problems. It covers the basics of digital logic through the lens of computer organization and programming. The reader should be able to design his or her own computer from the ground up at the end of the book. Logic simulation with Verilog is used throughout, assembly languages are introduced and discussed, and the fundamentals of computer architecture and embedded systems are touched upon, all in a cohesive design-driven framework suitable for class or self-study.
Fundamental concepts of mathematical modeling Modeling is one of the most effective, commonly used tools in engineering and the applied sciences. In this book, the authors deal with mathematical programming models both linear and nonlinear and across a wide range of practical applications. Whereas other books concentrate on standard methods of analysis, the authors focus on the power of modeling methods for solving practical problems–clearly showing the connection between physical and mathematical realities–while also describing and exploring the main concepts and tools at work. This highly computational coverage includes:
Building and Solving Mathematical Programming Models in Engineering and Science is practically suited for use as a professional reference for mathematicians, engineers, and applied or industrial scientists, while also tutorial and illustrative enough for advanced students in mathematics or engineering.
New Approaches to Circle Packing into the Square is devoted to the most recent results on the densest packing of equal circles in a square. In the last few decades, many articles have considered this question, which has been an object of interest since it is a hard challenge both in discrete geometry and in mathematical programming. The authors have studied this geometrical optimization problem for a long time, and they developed several new algorithms to solve it. The book completely covers the investigations on this topic.
This book provides a practical guide to applying soft-computing methods to interpret geophysical data. It discusses the design of neural networks with Matlab for geophysical data, as well as fuzzy logic and neuro-fuzzy concepts and their applications. In addition, it describes genetic algorithms for the automatic and/or intelligent processing and interpretation of geophysical data.
This volume honours the life and work of Solomon Feferman, one of the most prominent mathematical logicians of the latter half of the 20th century. In the collection of essays presented here, researchers examine Feferman's work on mathematical as well as specific methodological and philosophical issues that tie into mathematics. Feferman's work was largely based in mathematical logic (namely model theory, set theory, proof theory and computability theory), but also branched out into methodological and philosophical issues, making it well known beyond the borders of the mathematics community. With regard to methodological issues, Feferman supported concrete projects. On the one hand, these projects calibrate the proof theoretic strength of subsystems of analysis and set theory and provide ways of overcoming the limitations imposed by Goedel's incompleteness theorems through appropriate conceptual expansions. On the other, they seek to identify novel axiomatic foundations for mathematical practice, truth theories, and category theory. In his philosophical research, Feferman explored questions such as "What is logic?" and proposed particular positions regarding the foundations of mathematics including, for example, his "conceptual structuralism." The contributing authors of the volume examine all of the above issues. Their papers are accompanied by an autobiography presented by Feferman that reflects on the evolution and intellectual contexts of his work. The contributing authors critically examine Feferman's work and, in part, actively expand on his concrete mathematical projects. The volume illuminates Feferman's distinctive work and, in the process, provides an enlightening perspective on the foundations of mathematics and logic.
In a world of almost permanent and rapidly increasing electronic data availability, techniques of filtering, compressing, and interpreting this data to transform it into valuable and easily comprehensible information is of utmost importance. One key topic in this area is the capability to deduce future system behavior from a given data input. This book brings together for the first time the complete theory of data-based neurofuzzy modelling and the linguistic attributes of fuzzy logic in a single cohesive mathematical framework. After introducing the basic theory of data-based modelling, new concepts including extended additive and multiplicative submodels are developed and their extensions to state estimation and data fusion are derived. All these algorithms are illustrated with benchmark and real-life examples to demonstrate their efficiency. Chris Harris and his group have carried out pioneering work which has tied together the fields of neural networks and linguistic rule-based algortihms. This book is aimed at researchers and scientists in time series modeling, empirical data modeling, knowledge discovery, data mining, and data fusion.
The related fields of fractal image encoding and fractal image
analysis have blossomed in recent years. This book, originating
from a NATO Advanced Study Institute held in 1995, presents work by
leading researchers. It is developing the subjects at an
introductory level, but it also has some recent and exciting
results in both fields. |
![]() ![]() You may like...
Handbook of Electronic and Digital…
Thomas W. Leonhardt
Hardcover
Women's Fiction - A Guide to Popular…
Rebecca Vnuk, Nanette Donohue
Hardcover
R2,258
Discovery Miles 22 580
Storing and Transmitting Data - Rudolf…
Rudolf Ahlswede
Hardcover
|