![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > General theory of computing > Mathematical theory of computation
This open access book presents selected papers from International Symposium on Mathematics, Quantum Theory, and Cryptography (MQC), which was held on September 25-27, 2019 in Fukuoka, Japan. The international symposium MQC addresses the mathematics and quantum theory underlying secure modeling of the post quantum cryptography including e.g. mathematical study of the light-matter interaction models as well as quantum computing. The security of the most widely used RSA cryptosystem is based on the difficulty of factoring large integers. However, in 1994 Shor proposed a quantum polynomial time algorithm for factoring integers, and the RSA cryptosystem is no longer secure in the quantum computing model. This vulnerability has prompted research into post-quantum cryptography using alternative mathematical problems that are secure in the era of quantum computers. In this regard, the National Institute of Standards and Technology (NIST) began to standardize post-quantum cryptography in 2016. This book is suitable for postgraduate students in mathematics and computer science, as well as for experts in industry working on post-quantum cryptography.
This book provides a self-contained undergraduate course on quantum computing based on classroom-tested lecture notes. It reviews the fundamentals of quantum mechanics from the double-slit experiment to entanglement, before progressing to the basics of qubits, quantum gates, quantum circuits, quantum key distribution, and some of the famous quantum algorithms. As well as covering quantum gates in depth, it also describes promising platforms for their physical implementation, along with error correction, and topological quantum computing. With quantum computing expanding rapidly in the private sector, understanding quantum computing has never been so important for graduates entering the workplace or PhD programs. Assuming minimal background knowledge, this book is highly accessible, with rigorous step-by-step explanations of the principles behind quantum computation, further reading, and end-of-chapter exercises, ensuring that undergraduate students in physics and engineering emerge well prepared for the future.
The book is centered around the research areas of combinatorics, special functions, and computer algebra. What these research fields share is that many of their outstanding results do not only have applications in Mathematics, but also other disciplines, such as computer science, physics, chemistry, etc. A particular charm of these areas is how they interact and influence one another. For instance, combinatorial or special functions' techniques have motivated the development of new symbolic algorithms. In particular, first proofs of challenging problems in combinatorics and special functions were derived by making essential use of computer algebra. This book addresses these interdisciplinary aspects. Algorithmic aspects are emphasized and the corresponding software packages for concrete problem solving are introduced. Readers will range from graduate students, researchers to practitioners who are interested in solving concrete problems within mathematics and other research disciplines.
This book examines the mismatch between discrete programs, which lie at the center of modern applied mathematics, and the continuous space phenomena they simulate. The author considers whether we can imagine continuous spaces of programs, and asks what the structure of such spaces would be and how they would be constituted. He proposes a functional analysis of program spaces focused through the lens of iterative optimization. The author begins with the observation that optimization methods such as Genetic Algorithms, Evolution Strategies, and Particle Swarm Optimization can be analyzed as Estimation of Distributions Algorithms (EDAs) in that they can be formulated as conditional probability distributions. The probabilities themselves are mathematical objects that can be compared and operated on, and thus many methods in Evolutionary Computation can be placed in a shared vector space and analyzed using techniques of functional analysis. The core ideas of this book expand from that concept, eventually incorporating all iterative stochastic search methods, including gradient-based methods. Inspired by work on Randomized Search Heuristics, the author covers all iterative optimization methods and not just evolutionary methods. The No Free Lunch Theorem is viewed as a useful introduction to the broader field of analysis that comes from developing a shared mathematical space for optimization algorithms. The author brings in intuitions from several branches of mathematics such as topology, probability theory, and stochastic processes and provides substantial background material to make the work as self-contained as possible. The book will be valuable for researchers in the areas of global optimization, machine learning, evolutionary theory, and control theory.
This book on proof theory centers around the legacy of Kurt Schutte and its current impact on the subject. Schutte was the last doctoral student of David Hilbert who was the first to see that proofs can be viewed as structured mathematical objects amenable to investigation by mathematical methods (metamathematics). Schutte inaugurated the important paradigm shift from finite proofs to infinite proofs and developed the mathematical tools for their analysis. Infinitary proof theory flourished in his hands in the 1960s, culminating in the famous bound 0 for the limit of predicative mathematics (a fame shared with Feferman). Later his interests shifted to developing infinite proof calculi for impredicative theories. Schutte had a keen interest in advancing ordinal analysis to ever stronger theories and was still working on some of the strongest systems in his eighties. The articles in this volume from leading experts close to his research, show the enduring influence of his work in modern proof theory. They range from eye witness accounts of his scientific life to developments at the current research frontier, including papers by Schutte himself that have never been published before.
This book discusses the formalization of mathematical theories centering on complex analysis and matrix theory, covering topics such as algebraic systems, complex numbers, gauge integration, the Fourier transformation and its discrete counterpart, matrices and their transformation, inner product spaces, and function matrices. The formalization is performed using the interactive theorem prover HOL4, chiefly developed at the University of Cambridge. Many of the developments presented are now integral parts of the library of this prover. As mathematical developments continue to gain in complexity, sometimes demanding proofs of enormous sizes, formalization has proven to be invaluable in terms of obtaining real confidence in their correctness. This book provides a basis for the computer-aided verification of engineering systems constructed using the principles of complex analysis and matrix theory, as well as building blocks for the formalization of more involved mathematical theories.
Iceberg semantics is a new framework of Boolean semantics for mass nouns and count nouns in which the interpretation of a noun phrase rises up from a generating base and floats with its base on its Boolean part set, like an iceberg. The framework is shown to preserve the attractive features of classical Boolean semantics for count nouns; the book argues that Iceberg semantics forms a much better framework for studying mass nouns than the classical theory does. Iceberg semantics uses its notion of base to develop a semantic theory of the differences between mass nouns and count nouns and between different types of mass nouns, in particular between prototypical mass nouns (here called mess mass nouns) like water and mud versus object mass nouns (here called neat mass nouns) like poultry and pottery. The book shows in detail how and why neat mass nouns pattern semantically both with mess mass nouns and with count nouns. Iceberg semantics is a compositional theory and in Iceberg semantics the semantic distinctions defined apply to noun phrases of any complexity. The book studies in depth the semantics of classifier noun phrases (like three glasses of wine) and measure noun phrases (like three liters of wine). The classical wisdom is that classifier interpretations are count. Recent literature has argued compellingly that measure interpretations are mass. The book shows that both connections follow from the basic architecture of Iceberg semantics. Audience: Scholars and students in linguistics - in particular semantics, pragmatics, computational linguistics and syntax - and neighbouring disciplines like logic, philosophy of language, and cognitive science.
This book presents a collection of papers on recent advances in problems concerning dynamics, optimal control and optimization. In many chapters, computational techniques play a central role. Set-oriented techniques feature prominently throughout the book, yielding state-of-the-art algorithms for computing general invariant sets, constructing globally optimal controllers and solving multi-objective optimization problems.
This book presents a set theoretical development for the foundations of the theory of atomic and finitely supported structures. It analyzes whether a classical result can be adequately reformulated by replacing a 'non-atomic structure' with an 'atomic, finitely supported structure'. It also presents many specific properties, such as finiteness, cardinality, connectivity, fixed point, order and uniformity, of finitely supported atomic structures that do not have non-atomic correspondents. In the framework of finitely supported sets, the authors analyze the consistency of various forms of choice and related results. They introduce and study the notion of 'cardinality' by presenting various order and arithmetic properties. Finitely supported partially ordered sets, chain complete sets, lattices and Galois connections are studied, and new fixed point, calculability and approximation properties are presented. In this framework, the authors study the finitely supported L-fuzzy subsets of a finitely supported set and the finitely supported fuzzy subgroups of a finitely supported group. Several pairwise non-equivalent definitions for the notion of 'infinity' (Dedekind infinity, Mostowski infinity, Kuratowski infinity, Tarski infinity, ascending infinity) are introduced, compared and studied in the new framework. Relevant examples of sets that satisfy some forms of infinity while not satisfying others are provided. Uniformly supported sets are analyzed, and certain surprising properties are presented. Finally, some variations of the finite support requirement are discussed. The book will be of value to researchers in the foundations of set theory, algebra and logic.
This book details the conceptual foundations, design and implementation of the domain-specific language (DSL) development system DjDSL. DjDSL facilitates design-decision-making on and implementation of reusable DSL and DSL-product lines, and represents the state-of-the-art in language-based and composition-based DSL development. As such, it unites elements at the crossroads between software-language engineering, model-driven software engineering, and feature-oriented software engineering. The book is divided into six chapters. Chapter 1 ("DSL as Variable Software") explains the notion of DSL as variable software in greater detail and introduces readers to the idea of software-product line engineering for DSL-based software systems. Chapter 2 ("Variability Support in DSL Development") sheds light on a number of interrelated dimensions of DSL variability: variable development processes, variable design-decisions, and variability-implementation techniques for DSL. The three subsequent chapters are devoted to the key conceptual and technical contributions of DjDSL: Chapter 3 ("Variable Language Models") explains how to design and implement the abstract syntax of a DSL in a variable manner. Chapter 4 ("Variable Context Conditions") then provides the means to refine an abstract syntax (language model) by using composable context conditions (invariants). Next, Chapter 5 ("Variable Textual Syntaxes") details solutions to implementing variable textual syntaxes for different types of DSL. In closing, Chapter 6 ("A Story of a DSL Family") shows how to develop a mixed DSL in a step-by-step manner, demonstrating how the previously introduced techniques can be employed in an advanced example of developing a DSL family. The book is intended for readers interested in language-oriented as well as model-driven software development, including software-engineering researchers and advanced software developers alike. An understanding of software-engineering basics (architecture, design, implementation, testing) and software patterns is essential. Readers should especially be familiar with the basics of object-oriented modelling (UML, MOF, Ecore) and programming (e.g., Java).
This book provides detailed descriptions of big data solutions for activity detection and forecasting of very large numbers of moving entities spread across large geographical areas. It presents state-of-the-art methods for processing, managing, detecting and predicting trajectories and important events related to moving entities, together with advanced visual analytics methods, over multiple heterogeneous, voluminous, fluctuating and noisy data streams from moving entities, correlating them with data from archived data sources expressing e.g. entities' characteristics, geographical information, mobility patterns, mobility regulations and intentional data. The book is divided into six parts: Part I discusses the motivation and background of mobility forecasting supported by trajectory-oriented analytics, and includes specific problems and challenges in the aviation (air-traffic management) and the maritime domains. Part II focuses on big data quality assessment and processing, and presents novel technologies suitable for mobility analytics components. Next, Part III describes solutions toward processing and managing big spatio-temporal data, particularly enriching data streams and integrating streamed and archival data to provide coherent views of mobility, and storing of integrated mobility data in large distributed knowledge graphs for efficient query-answering. Part IV focuses on mobility analytics methods exploiting (online) processed, synopsized and enriched data streams as well as (offline) integrated, archived mobility data, and highlights future location and trajectory prediction methods, distinguishing between short-term and more challenging long-term predictions. Part V examines how methods addressing data management, data processing and mobility analytics are integrated in big data architectures with distinctive characteristics compared to other known big data paradigmatic architectures. Lastly, Part VI covers important ethical issues that research on mobility analytics should address. Providing novel approaches and methodologies related to mobility detection and forecasting needs based on big data exploration, processing, storage, and analysis, this book will appeal to computer scientists and stakeholders in various application domains.
This text presents an algebraic approach to the construction of several important families of quantum codes derived from classical codes by applying the well-known Calderbank-Shor-Steane (CSS), Hermitian, and Steane enlargement constructions to certain classes of classical codes. In addition, the book presents families of asymmetric quantum codes with good parameters and provides a detailed description of the procedures adopted to construct families of asymmetric quantum convolutional codes.Featuring accessible language and clear explanations, the book is suitable for use in advanced undergraduate and graduate courses as well as for self-guided study and reference. It provides an expert introduction to algebraic techniques of code construction and, because all of the constructions are performed algebraically, it enables the reader to construct families of codes, rather than only codes with specific parameters. The text offers an abundance of worked examples, exercises, and open-ended problems to motivate the reader to further investigate this rich area of inquiry. End-of-chapter summaries and a glossary of key terms allow for easy review and reference.
This book gathers the proceedings of the 2018 Abel Symposium, which was held in Geiranger, Norway, on June 4-8, 2018. The symposium offered an overview of the emerging field of "Topological Data Analysis". This volume presents papers on various research directions, notably including applications in neuroscience, materials science, cancer biology, and immune response. Providing an essential snapshot of the status quo, it represents a valuable asset for practitioners and those considering entering the field.
Keeping students involved and actively learning is challenging. Instructors in computer science are aware of the cognitive value of modelling puzzles and often use logical puzzles as an efficient pedagogical instrument to engage students and develop problem-solving skills. This unique book is a comprehensive resource that offers teachers and students fun activities to teach and learn logic. It provides new, complete, and running formalisation in Propositional and First Order Logic for over 130 logical puzzles, including Sudoku-like puzzles, zebra-like puzzles, island of truth, lady and tigers, grid puzzles, strange numbers, or self-reference puzzles. Solving puzzles with theorem provers can be an effective cognitive incentive to motivate students to learn logic. They will find a ready-to-use format which illustrates how to model each puzzle, provides running implementations, and explains each solution. This concise and easy-to-follow textbook is a much-needed support tool for students willing to explore beyond the introductory level of learning logic and lecturers looking for examples to heighten student engagement in their computer science courses.
This book explores the impact of nonlinearity on a broad range of areas, including time-honored fields such as biology, geometry, and topology, but also modern ones such as quantum mechanics, networks, metamaterials and artificial intelligence. The concept of nonlinearity is a universal feature in mathematics, physics, chemistry and biology, and is used to characterize systems whose behavior does not amount to a superposition of simple building blocks, but rather features complex and often chaotic patterns and phenomena. Each chapter of the book features a synopsis that not only recaps the recent progress in each field but also charts the challenges that lie ahead. This interdisciplinary book presents contributions from a diverse group of experts from various fields to provide an overview of each field's past, present and future. It will appeal to both beginners and seasoned researchers in nonlinear science, numerous areas of physics (optics, quantum physics, biophysics), and applied mathematics (ODEs, PDEs, dynamical systems, machine learning) as well as engineering.
This book introduces the reader with little or no previous computer-programming experience to the Python programming language of interest for a physicist or a natural-sciences student. The book starts with basic interactive Python in order to acquire an introductory familiarity with the language, than tackle Python scripts (programs) of increasing complexity, that the reader is invited to run on her/his computer. All program listings are discussed in detail, and the reader is invited to experiment on what happens if some code lines are modified. The reader is introduced to Matplotlib graphics for the generation of figures representing data and function plots and, for instance, field lines. Animated function plots are also considered. A chapter is dedicated to the numerical solution of algebraic and transcendental equations, the basic mathematical principles are discussed and the available Python tools for the solution are presented. A further chapter is dedicated to the numerical solution of ordinary differential equations. This is of vital importance for the physicist, since differential equations are at the base of both classical physics (Newton's equations) and quantum mechanics (Schroedinger's equation). The shooting method for the numerical solution of ordinary differential equations with boundary conditions at two boundaries is also presented. Python programs for the solution of two quantum-mechanics problems are discussed as examples. Two chapters are dedicated to Tkinter graphics, which gives the user more freedom than Matplotlib, and to Tkinter animation. Programs displaying the animation of physical problems involving the solution of ordinary differential equations (for which in most cases there is no algebraic solution) in real time are presented and discussed. Finally, 3D animation is presented with Vpython.
This book on optimization includes forewords by Michael I. Jordan, Zongben Xu and Zhi-Quan Luo. Machine learning relies heavily on optimization to solve problems with its learning models, and first-order optimization algorithms are the mainstream approaches. The acceleration of first-order optimization algorithms is crucial for the efficiency of machine learning. Written by leading experts in the field, this book provides a comprehensive introduction to, and state-of-the-art review of accelerated first-order optimization algorithms for machine learning. It discusses a variety of methods, including deterministic and stochastic algorithms, where the algorithms can be synchronous or asynchronous, for unconstrained and constrained problems, which can be convex or non-convex. Offering a rich blend of ideas, theories and proofs, the book is up-to-date and self-contained. It is an excellent reference resource for users who are seeking faster optimization algorithms, as well as for graduate students and researchers wanting to grasp the frontiers of optimization in machine learning in a short time.
This is the first textbook that allows readers who may be unfamiliar with matrices to understand a variety of multivariate analysis procedures in matrix forms. By explaining which models underlie particular procedures and what objective function is optimized to fit the model to the data, it enables readers to rapidly comprehend multivariate data analysis. Arranged so that readers can intuitively grasp the purposes for which multivariate analysis procedures are used, the book also offers clear explanations of those purposes, with numerical examples preceding the mathematical descriptions. Supporting the modern matrix formulations by highlighting singular value decomposition among theorems in matrix algebra, this book is useful for undergraduate students who have already learned introductory statistics, as well as for graduate students and researchers who are not familiar with matrix-intensive formulations of multivariate data analysis. The book begins by explaining fundamental matrix operations and the matrix expressions of elementary statistics. Then, it offers an introduction to popular multivariate procedures, with each chapter featuring increasing advanced levels of matrix algebra. Further the book includes in six chapters on advanced procedures, covering advanced matrix operations and recently proposed multivariate procedures, such as sparse estimation, together with a clear explication of the differences between principal components and factor analyses solutions. In a nutshell, this book allows readers to gain an understanding of the latest developments in multivariate data science.
This book features selected papers presented at the 2nd International Conference on Advanced Computing Technologies and Applications, held at SVKM's Dwarkadas J. Sanghvi College of Engineering, Mumbai, India, from 28 to 29 February 2020. Covering recent advances in next-generation computing, the book focuses on recent developments in intelligent computing, such as linguistic computing, statistical computing, data computing and ambient applications.
This book covers applications of machine learning in artificial intelligence. The specific topics covered include human language, heterogeneous and streaming data, unmanned systems, neural information processing, marketing and the social sciences, bioinformatics and robotics, etc. It also provides a broad range of techniques that can be successfully applied and adopted in different areas. Accordingly, the book offers an interesting and insightful read for scholars in the areas of computer vision, speech recognition, healthcare, business, marketing, and bioinformatics.
This book analyzes techniques that use the direct and inverse fuzzy transform for image processing and data analysis. The book is divided into two parts, the first of which describes methods and techniques that use the bi-dimensional fuzzy transform method in image analysis. In turn, the second describes approaches that use the multidimensional fuzzy transform method in data analysis. An F-transform in one variable is defined as an operator which transforms a continuous function f on the real interval [a,b] in an n-dimensional vector by using n-assigned fuzzy sets A1, ... , An which constitute a fuzzy partition of [a,b]. Then, an inverse F-transform is defined in order to convert the n-dimensional vector output in a continuous function that equals f up to an arbitrary quantity . We may limit this concept to the finite case by defining the discrete F-transform of a function f in one variable, even if it is not known a priori. A simple extension of this concept to functions in two variables allows it to be used for the coding/decoding and processing of images. Moreover, an extended version with multidimensional functions can be used to address a host of topics in data analysis, including the analysis of large and very large datasets. Over the past decade, many researchers have proposed applications of fuzzy transform techniques for various image processing topics, such as image coding/decoding, image reduction, image segmentation, image watermarking and image fusion; and for such data analysis problems as regression analysis, classification, association rule extraction, time series analysis, forecasting, and spatial data analysis. The robustness, ease of use, and low computational complexity of fuzzy transforms make them a powerful fuzzy approximation tool suitable for many computer science applications. This book presents methods and techniques based on the use of fuzzy transforms in various applications of image processing and data analysis, including image segmentation, image tamper detection, forecasting, and classification, highlighting the benefits they offer compared with traditional methods. Emphasis is placed on applications of fuzzy transforms to innovative problems, such as massive data mining, and image and video security in social networks based on the application of advanced fragile watermarking systems. This book is aimed at researchers, students, computer scientists and IT developers to acquire the knowledge and skills necessary to apply and implement fuzzy transforms-based techniques in image and data analysis applications.
This book constitutes the proceedings of the 23rd International Conference on Developments in Language Theory, DLT 2019, held in Warsaw, Poland, in August 2019. The 20 full papers presented together with three invited talks were carefully reviewed and selected from 30 submissions. The papers cover the following topics and areas: combinatorial and algebraic properties of words and languages; grammars, acceptors and transducers for strings, trees, graphics, arrays; algebraic theories for automata and languages; codes; efficient text algorithms; symbolic dynamics; decision problems; relationships to complexity theory and logic; picture description and analysis, polyominoes and bidimensional patterns; cryptography; concurrency; celluar automata; bio-inspired computing; quantum computing.
This book constitutes the refereed proceedings of the 22nd International Conference on Theory and Applications of Satisfiability Testing, SAT 2019, held in Lisbon, Portugal, UK, in July 2019. The 19 revised full papers presented together with 7 short papers were carefully reviewed and selected from 64 submissions. The papers address different aspects of SAT interpreted in a broad sense, including (but not restricted to) theoretical advances (such as exact algorithms, proof complexity, and other complexity issues), practical search algorithms, knowledge compilation, implementation-level details of SAT solvers and SAT-based systems, problem encodings and reformulations, applications (including both novel application domains and improvements to existing approaches), as well as case studies and reports on findings based on rigorous experimentation.
This book constitutes the refereed proceedings of the 20th International Conference on Data Analytics and Management in Data Intensive Domains, DAMDID/RCDL 2018, held in Moscow, Russia, in October 2018.The 9 revised full papers presented together with three invited papers were carefully reviewed and selected from 54 submissions. The papers are organized in the following topical sections: FAIR data infrastructures, interoperability and reuse; knowledge representation; data models; data analysis in astronomy; text search and processing; distributed computing; information extraction from text.
This book covers different, current research directions in the context of variational methods for non-linear geometric data. Each chapter is authored by leading experts in the respective discipline and provides an introduction, an overview and a description of the current state of the art. Non-linear geometric data arises in various applications in science and engineering. Examples of nonlinear data spaces are diverse and include, for instance, nonlinear spaces of matrices, spaces of curves, shapes as well as manifolds of probability measures. Applications can be found in biology, medicine, product engineering, geography and computer vision for instance. Variational methods on the other hand have evolved to being amongst the most powerful tools for applied mathematics. They involve techniques from various branches of mathematics such as statistics, modeling, optimization, numerical mathematics and analysis. The vast majority of research on variational methods, however, is focused on data in linear spaces. Variational methods for non-linear data is currently an emerging research topic. As a result, and since such methods involve various branches of mathematics, there is a plethora of different, recent approaches dealing with different aspects of variational methods for nonlinear geometric data. Research results are rather scattered and appear in journals of different mathematical communities. The main purpose of the book is to account for that by providing, for the first time, a comprehensive collection of different research directions and existing approaches in this context. It is organized in a way that leading researchers from the different fields provide an introductory overview of recent research directions in their respective discipline. As such, the book is a unique reference work for both newcomers in the field of variational methods for non-linear geometric data, as well as for established experts that aim at to exploit new research directions or collaborations. Chapter 9 of this book is available open access under a CC BY 4.0 license at link.springer.com. |
![]() ![]() You may like...
The Larger Illustrated Guide Sasol Birds…
Ian Sinclair, Phil Hockey
Paperback
She Surf - The Rise of Female Surfing
Lauren L Hill, Gestalten
Hardcover
Purchasing and supply management
J.A. Badenhorst-Weiss, J.O. Cilliers, …
Paperback
![]() R821 Discovery Miles 8 210
Ecclesia Semper Reformanda Est / The…
David G. Barker, Michael A.G. Haykin, …
Hardcover
Americathon - The Skits Behind the…
Phil Proctor, Peter Bergman
Hardcover
R657
Discovery Miles 6 570
|