Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > General theory of computing > Mathematical theory of computation
Research and development in the pioneering field of quantum computing involve just about every facet of science and engineering, including the significant areas of mathematics and physics. Based on the firm understanding that mathematics and physics are equal partners in the continuing study of quantum science, Mathematics of Quantum Computation and Quantum Technology explores the rapid mathematical advancements made in this field in recent years. Novel Viewpoints on Numerous Aspects of Quantum Computing and Technology Edited by a well-respected team of experts, this volume compiles contributions from specialists across various disciplines. It contains four main parts, beginning with topics in quantum computing that include quantum algorithms and hidden subgroups, quantum search, algorithmic complexity, and quantum simulation. The next section covers quantum technology, such as mathematical tools, quantum wave functions, superconducting quantum computing interference devices (SQUIDs), and optical quantum computing. The section on quantum information deals with error correction, cryptography, entanglement, and communication. The final part explores topological quantum computation, knot theory, category algebra, and logic. The Tools You Need to Tackle the Next Generation of Quantum Technology This book facilitates both the construction of a common quantum language and the development of interdisciplinary quantum techniques, which will aid efforts in the pursuit of the ultimate goal-a "real" scalable quantum computer.
Structured methodologies are a popular and powerful tool in information systems development. Many different ones exist, each employing a number of models and so a specification must be converted from one form to another during the development process. To solve this problem, Dr Tse proposes in this 1991 book a unifying framework behind popular structured models. He approaches the problem from the viewpoints of algebra and category theory. He not only develops the frameworks but also illustrates their practical and theoretical usefulness. Thus this book will provide insight for software engineers into how methodologies can be formalised and will open up a range of applications and problems for theoretical computer scientists.
Logic programming was based on first-order logic. Higher-order logics can also lead to theories of theorem-proving. This book introduces just such a theory, based on a lambda-calculus formulation of a clausal logic with equality, known as the Clausal Theory of Types. By restricting this logic to Horn clauses, a concise form of logic programming that incorporates functional programming is achieved. The book begins by reviewing the fundamental Skolem-Herbrand-Goedel Theorem and resolution, which are then extrapolated to a higher-order setting; this requires introducing higher-order equational unification which builds in higher-order equational theories and uses higher-order rewriting. The logic programming language derived has the unique property of being sound and complete with respect to Henkin-Andrews general models, and consequently of treating equivalent terms as identical. First published in 1993, the book can be used for graduate courses in theorem-proving, but will be of interest to all working in declarative programming.
Constraint Handling Rules (CHR) is both a theoretical formalism based on logic and a practical programming language based on rules. This book, written by the creator of CHR, describes the theory of CHR and how to use it in practice. It is supported by a website containing teaching materials, online demos, and free downloads of the language. After a basic tutorial, the author describes in detail the CHR language and discusses guaranteed properties of CHR programs. The author then compares CHR with other formalisms and languages and illustrates how it can capture their essential features. Finally, larger programs are introduced and analyzed in detail. The book is ideal for graduate students and lecturers, and for more experienced programmers and researchers, who can use it for self-study. Exercises with selected solutions, and bibliographic remarks are included at the ends of chapters. The book is the definitive reference on the subject.
This 1993 book shows how formal logic can be used to specify the behaviour of hardware designs and reason about their correctness. A primary theme of the book is the use of abstraction in hardware specification and verification. The author describes how certain fundamental abstraction mechanisms for hardware verification can be formalised in logic and used to express assertions about design correctness and the relative accuracy of models of hardware behaviour. His approach is pragmatic and driven by examples. He also includes an introduction to higher-order logic, which is a widely used formalism in this subject, and describes how that formalism is actually used for hardware verification. The book is based in part on the author's own research as well as on graduate teaching. Thus it can be used to accompany courses on hardware verification and as a resource for research workers.
Originally published in 1988, this book presents an introduction to lambda-calculus and combinators without getting lost in the details of mathematical aspects of their theory. Lambda-calculus is treated here as a functional language and its relevance to computer science is clearly demonstrated. The main purpose of the book is to provide computer science students and researchers with a firm background in lambda-calculus and combinators and show the applicabillity of these theories to functional programming. The presentation of the material is self-contained. It can be used as a primary text for a course on functional programming. It can also be used as a supplementary text for courses on the structure and implementation of programming languages, theory of computing, or semantics of programming languages.
Randomized algorithms have become a central part of the algorithms curriculum, based on their increasingly widespread use in modern applications. This book presents a coherent and unified treatment of probabilistic techniques for obtaining high probability estimates on the performance of randomized algorithms. It covers the basic toolkit from the Chernoff-Hoeffding bounds to more sophisticated techniques like martingales and isoperimetric inequalities, as well as some recent developments like Talagrand's inequality, transportation cost inequalities and log-Sobolev inequalities. Along the way, variations on the basic theme are examined, such as Chernoff-Hoeffding bounds in dependent settings. The authors emphasise comparative study of the different methods, highlighting respective strengths and weaknesses in concrete example applications. The exposition is tailored to discrete settings sufficient for the analysis of algorithms, avoiding unnecessary measure-theoretic details, thus making the book accessible to computer scientists as well as probabilists and discrete mathematicians.
This book is a thorough introduction to the formal foundations and practical applications of Bayesian networks. It provides an extensive discussion of techniques for building Bayesian networks that model real-world situations, including techniques for synthesizing models from design, learning models from data, and debugging models using sensitivity analysis. It also treats exact and approximate inference algorithms at both theoretical and practical levels. The treatment of exact algorithms covers the main inference paradigms based on elimination and conditioning and includes advanced methods for compiling Bayesian networks, time-space tradeoffs, and exploiting local structure of massively connected networks. The treatment of approximate algorithms covers the main inference paradigms based on sampling and optimization and includes influential algorithms such as importance sampling, MCMC, and belief propagation. The author assumes very little background on the covered subjects, supplying in-depth discussions for theoretically inclined readers and enough practical details to provide an algorithmic cookbook for the system developer.
This textbook aims to help the reader develop an in-depth understanding of logical reasoning and gain knowledge of the theory of computation. The book combines theoretical teaching and practical exercises; the latter is realised in Isabelle/HOL, a modern theorem prover, and PAT, an industry-scale model checker. I also give entry-level tutorials on the two software to help the reader get started. By the end of the book, the reader should be proficient in both software. Content-wise, this book focuses on the syntax, semantics and proof theory of various logics; automata theory, formal languages, computability and complexity. The final chapter closes the gap with a discussion on the insight that links logic with computation. This book is written for a high-level undergraduate course or a Master's course. The hybrid skill set of practical theorem proving and model checking should be helpful for the future of readers should they pursue a research career or engineering in formal methods.
First published in 1998, this textbook is a broad but rigourous survey of the theoretical basis for the design, definition and implementation of programming languages and of systems for specifying and proving programme behaviour. Both imperative and functional programming are covered, as well as the ways of integrating these aspects into more general languages. Recognising a unity of technique beneath the diversity of research in programming languages, the author presents an integrated treatment of the basic principles of the subject. He identifies the relatively small number of concepts, such as compositional semantics, binding structure, domains, transition systems and inference rules, that serve as the foundation of the field. Assuming only knowledge of elementary programming and mathematics, this text is perfect for advanced undergraduate and beginning graduate courses in programming language theory and also will appeal to researchers and professionals in designing or implementing computer languages.
The goal of this book is to provide a comprehensive and systematic introduction to the important and highly applicable method of data refinement and the simulation methods used for proving its correctness. The authors concentrate in the first part on the general principles needed to prove data refinement correct. They begin with an explanation of the fundamental notions, showing that data refinement proofs reduce to proving simulation. The topics of Hoare Logic and the Refinement Calculus are introduced and a general theory of simulations is developed and related to them. Accessibility and comprehension are emphasized in order to guide newcomers to the area. The book's second part contains a detailed survey of important methods in this field, such as VDM, and the methods due to Abadi & Lamport, Hehner, Lynch and Reynolds, Back's refinement calculus and Z. All these methods are carefully analysed, and shown to be either imcomplete, with counterexamples to their application, or to be always applicable whenever data refinement holds. This is shown by proving, for the first time, that all these methods can be described and analyzed in terms of two simple notions: forward and backward simulation. The book is self-contained, going from advanced undergraduate level and taking the reader to the state of the art in methods for proving simulation.
Modern computer networks now circle the world, but the transmission of information between them depends on the many different protocols that define the behaviour of the sender and receiver. It is clear therefore, that the accurate description of these protocols is important if harmonious communication is to be maintained. In this book the authors use the formal specification language PSF to provide an unambiguous description of several communication protocols of varying levels of complexity, ranging from the alternating bit protocol to the token ring protocol. Beginners, as well as professionals in the field of communication protocols, will benefit from both the methods of specification described and the protocols discussed in this book.
Number theory and algebra play an increasingly significant role in computing and communications, as evidenced by the striking applications of these subjects to such fields as cryptography and coding theory. This introductory book emphasizes algorithms and applications, such as cryptography and error correcting codes, and is accessible to a broad audience. The presentation alternates between theory and applications in order to motivate and illustrate the mathematics. The mathematical coverage includes the basics of number theory, abstract algebra and discrete probability theory. This edition now includes over 150 new exercises, ranging from the routine to the challenging, that flesh out the material presented in the body of the text, and which further develop the theory and present new applications. The material has also been reorganized to improve clarity of exposition and presentation. Ideal as a textbook for introductory courses in number theory and algebra, especially those geared towards computer science students.
Combinatory logic and lambda-calculus, originally devised in the 1920's, have since developed into linguistic tools, especially useful in programming languages. The authors' previous book served as the main reference for introductory courses on lambda-calculus for over 20 years: this long-awaited new version is thoroughly revised and offers a fully up-to-date account of the subject, with the same authoritative exposition. The grammar and basic properties of both combinatory logic and lambda-calculus are discussed, followed by an introduction to type-theory. Typed and untyped versions of the systems, and their differences, are covered. Lambda-calculus models, which lie behind much of the semantics of programming languages, are also explained in depth. The treatment is as non-technical as possible, with the main ideas emphasized and illustrated by examples. Many exercises are included, from routine to advanced, with solutions to most at the end of the book.
Domain theory is an established part of theoretical computer science, used in giving semantics to programming languages and logics. In mathematics and logic it has also proved to be useful in the study of algorithms. This book is devoted to providing a unified and self-contained treatment of the subject. The theory is presented in a mathematically precise manner which nevertheless is accessible to mathematicians and computer scientists alike. The authors begin with the basic theory including domain equations, various domain representations and universal domains. They then proceed to more specialized topics such as effective and power domains, models of lambda-calculus and so on. In particular, the connections with ultrametric spaces and the Kleene-Kreisel continuous functionals are made precise. Consequently the text will be useful as an introductory textbook (earlier versions have been class-tested in Uppsala, Gothenburg, Passau, Munich and Swansea), or as a general reference for professionals in computer science and logic.
This book describes the mathematical aspects of the semantics of programming languages. The main goals are to provide formal tools to assess the meaning of programming constructs in both a language-independent and a machine-independent way and to prove properties about programs, such as whether they terminate, or whether their result is a solution of the problem they are supposed to solve. In order to achieve this the authors first present, in an elementary and unified way, the theory of certain topological spaces that have proved of use in the modeling of various families of typed lambda calculi considered as core programming languages and as meta-languages for denotational semantics. This theory is now known as Domain Theory, and was founded as a subject by Scott and Plotkin. One of the main concerns is to establish links between mathematical structures and more syntactic approaches to semantics, often referred to as operational semantics, which is also described. This dual approach has the double advantage of motivating computer scientists to do some mathematics and of interesting mathematicians in unfamiliar application areas from computer science.
Complexity theory is a central field of the theoretical foundations of computer science. It is concerned with the general study of the intrinsic complexity of computational tasks; that is, it addresses the question of what can be achieved within limited time (and/or with other limited natural computational resources). This book offers a conceptual perspective on complexity theory. It is intended to serve as an introduction for advanced undergraduate and graduate students, either as a textbook or for self-study. The book will also be useful to experts, since it provides expositions of the various sub-areas of complexity theory such as hardness amplification, pseudorandomness and probabilistic proof systems. In each case, the author starts by posing the intuitive questions that are addressed by the sub-area and then discusses the choices made in the actual formulation of these questions, the approaches that lead to the answers, and the ideas that are embedded in these answers.
This new edition of a well-received textbook provides a concise introduction to both the theoretical and experimental aspects of quantum information at the graduate level. While the previous edition focused on theory, the book now incorporates discussions of experimental platforms. Several chapters on experimental implementations of quantum information protocols have been added: implementations using neutral atoms, trapped ions, optics, and solidstate systems are each presented in its own chapter. Previous chapters on entanglement, quantum measurements, quantum dynamics, quantum cryptography, and quantum algorithms have been thoroughly updated, and new additions include chapters on the stabilizer formalism and the Gottesman-Knill theorem as well as aspects of classical and quantum information theory. To facilitate learning, each chapter starts with a clear motivation to the topic and closes with exercises and a recommended reading list. Quantum Information Processing: Theory and Implementation will be essential to graduate students studying quantum information as well as and researchers in other areas of physics who wish to gain knowledge in the field.
The Z notation is a language for expressing mathematical specifications of computing systems. By providing a formal semantics for Z, this book justifies the claim that Z is a precise specification language, and provides a standard framework for understanding Z specifications. Z is compared with other methods such as VDM and CLEAR, and in the final chapter, a number of studies in Z style are presented to show that Z can be used in a wide variety of specification tasks. This book will appeal to those who wish to deepen their understanding of the mathematical background of the Z method, those who want a concise definition of the language, and those interested in the design of specification languages.
The lecture courses in this work are derived from the SERC 'Logic for IT' Summer School and Conference on Proof Theory held at Leeds University. The contributions come from acknowledged experts and comprise expository and research articles; put together in this book they form an invaluable introduction to proof theory that is aimed at both mathematicians and computer scientists.
Type theory is one of the most important tools in the design of higher-level programming languages, such as ML. This book introduces and teaches its techniques by focusing on one particularly neat system and studying it in detail. By concentrating on the principles that make the theory work in practice, the author covers all the key ideas without getting involved in the complications of more advanced systems. This book takes a type-assignment approach to type theory, and the system considered is the simplest polymorphic one. The author covers all the basic ideas, including the system's relation to propositional logic, and gives a careful treatment of the type-checking algorithm that lies at the heart of every such system. Also featured are two other interesting algorithms that until now have been buried in inaccessible technical literature. The mathematical presentation is rigorous but clear, making it the first book at this level that can be used as an introduction to type theory for computer scientists.
This collection of peer-reviewed workshop papers provides comprehensive coverage of cutting-edge research into topological approaches to data analysis and visualization. It encompasses the full range of new algorithms and insights, including fast homology computation, comparative analysis of simplification techniques, and key applications in materials and medical science. The book also addresses core research challenges such as the representation of large and complex datasets, and integrating numerical methods with robust combinatorial algorithms. In keeping with the focus of the TopoInVis 2017 Workshop, the contributions reflect the latest advances in finding experimental solutions to open problems in the sector. They provide an essential snapshot of state-of-the-art research, helping researchers to keep abreast of the latest developments and providing a basis for future work. Gathering papers by some of the world's leading experts on topological techniques, the book represents a valuable contribution to a field of growing importance, with applications in disciplines ranging from engineering to medicine.
Formal Methods in Computer Science gives students a comprehensive introduction to formal methods and their application in software and hardware specification and verification. The first part introduces some fundamentals in formal methods, including set theory, functions, finite state machines, and regular expressions. The second part focuses on logic, a powerful formal language in specifying systems properties. It covers propositional logic, predicate logic, temporal logic, and model checking. The third part presents Petri nets, the most popular formal language in system behavior modeling. In additional to regular Petri nets, this part also examines timed Petri nets and high-level Petri nets. The textbook is ideal for undergraduate or graduate courses in computer engineering, software engineering, computer science, and information technology programs. Parts of the book are useful reading material in undergraduate computer course and as a reference guide for students researching the area of formal system specification and validation. Features * Introduces a comprehensive, yet manageable set of formal techniques for computer science students * Stresses real-world application value of each formal technique * Offers a good set of exercises which help students better understand the presented techniques * Also offers a prepared source code for downloading and non-commercial use
Assuming little previous mathematical knowledge, Error Correcting Codes provides a sound introduction to key areas of the subject. Topics have been chosen for their importance and practical significance, which Baylis demonstrates in a rigorous but gentle mathematical style.
This study explores an approach to text generation that interprets systemic grammar as a computational representation. Terry Patten demonstrates that systemic grammar can be easily and automatically translated into current AI knowledge representations and efficiently processed by the same knowledge-based techniques currently exploited by expert systems. Thus the fundamental methodological problem of interfacing specialized computational representations with equally specialized linguistic representations can be resolved. The study provides a detailed discussion of a substantial implementation involving a relatively large systemic grammar, and a formal model of the method. It represents a fundamental and productive contribution to the literature on text generation. |
You may like...
AI-based Metaheuristics for Information…
Apoorva S Shastri, Mangal Singh, …
Hardcover
R2,371
Discovery Miles 23 710
Computational Geometry - Algorithms and…
Mark De Berg, Otfried Cheong, …
Paperback
R1,625
Discovery Miles 16 250
Mathematical Aspects of Deep Learning
Philipp Grohs, Gitta Kutyniok
Hardcover
The Less Is More Linear Algebra of…
Daniela Calvetti, Erkki Somersalo
Paperback
R1,866
Discovery Miles 18 660
PCI Compliance - Understand and…
Branden Williams, James Adamson
Paperback
R1,370
Discovery Miles 13 700
|