![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Science & Mathematics > Mathematics > Mathematical foundations > Mathematical logic
Quantitative thinking is our inclination to view natural and everyday phenomena through a lens of measurable events, with forecasts, odds, predictions, and likelihood playing a dominant part. The Error of Truth recounts the astonishing and unexpected tale of how quantitative thinking came to be, and its rise to primacy in the nineteenth and early twentieth centuries. Additionally, it considers how seeing the world through a quantitative lens has shaped our perception of the world we live in, and explores the lives of the individuals behind its early establishment. This worldview was unlike anything humankind had before, and it came about because of a momentous human achievement: we had learned how to measure uncertainty. Probability as a science was conceptualised. As a result of probability theory, we now had correlations, reliable predictions, regressions, the bellshaped curve for studying social phenomena, and the psychometrics of educational testing. Significantly, these developments happened during a relatively short period in world history- roughly, the 130-year period from 1790 to 1920, from about the close of the Napoleonic era, through the Enlightenment and the Industrial Revolutions, to the end of World War I. At which time, transportation had advanced rapidly, due to the invention of the steam engine, and literacy rates had increased exponentially. This brief period in time was ready for fresh intellectual activity, and it gave a kind of impetus for the probability inventions. Quantification is now everywhere in our daily lives, such as in the ubiquitous microchip in smartphones, cars, and appliances; in the Bayesian logic of artificial intelligence, as well as applications in business, engineering, medicine, economics, and elsewhere. Probability is the foundation of quantitative thinking. The Error of Truth tells its story- when, why, and how it happened.
This book proves some important new theorems in the theory of canonical inner models for large cardinal hypotheses, a topic of central importance in modern set theory. In particular, the author 'completes' the theory of Fine Structure and Iteration Trees (FSIT) by proving a comparison theorem for mouse pairs parallel to the FSIT comparison theorem for pure extender mice, and then using the underlying comparison process to develop a fine structure theory for strategy mice. Great effort has been taken to make the book accessible to non-experts so that it may also serve as an introduction to the higher reaches of inner model theory. It contains a good deal of background material, some of it unpublished folklore, and includes many references to the literature to guide further reading. An introductory essay serves to place the new results in their broader context. This is a landmark work in inner model theory that should be in every set theorist's library.
Computing in Nonlinear Media and Automata Collectives presents an account of new ways to design massively parallel computing devices in advanced mathematical models, such as cellular automata and lattice swarms, from unconventional materials, including chemical solutions, bio-polymers, and excitable media.
Computability theory is a branch of mathematical logic and computer science that has become increasingly relevant in recent years. The field has developed growing connections in diverse areas of mathematics, with applications in topology, group theory, and other subfields. In A Hierarchy of Turing Degrees, Rod Downey and Noam Greenberg introduce a new hierarchy that allows them to classify the combinatorics of constructions from many areas of computability theory, including algorithmic randomness, Turing degrees, effectively closed sets, and effective structure theory. This unifying hierarchy gives rise to new natural definability results for Turing degree classes, demonstrating how dynamic constructions become reflected in definability. Downey and Greenberg present numerous construction techniques involving high-level nonuniform arguments, and their self-contained work is appropriate for graduate students and researchers. Blending traditional and modern research results in computability theory, A Hierarchy of Turing Degrees establishes novel directions in the field.
In these essays Geoffrey Hellman presents a strong case for a healthy pluralism in mathematics and its logics, supporting peaceful coexistence despite what appear to be contradictions between different systems, and positing different frameworks serving different legitimate purposes. The essays refine and extend Hellman's modal-structuralist account of mathematics, developing a height-potentialist view of higher set theory which recognizes indefinite extendability of models and stages at which sets occur. In the first of three new essays written for this volume, Hellman shows how extendability can be deployed to derive the axiom of Infinity and that of Replacement, improving on earlier accounts; he also shows how extendability leads to attractive, novel resolutions of the set-theoretic paradoxes. Other essays explore advantages and limitations of restrictive systems - nominalist, predicativist, and constructivist. Also included are two essays, with Solomon Feferman, on predicative foundations of arithmetic.
This story of a highly intelligent observer of the turbulent 20th century who was intimately involved as the secretary and bodyguard to Leon Trotsky is based on extensive interviews with the subject, Jean van Heijenoort, and his family, friends, and colleagues. The author has captured the personal drama and the professional life of her protagonist--ranging from the political passion of a young intellectual to the scientific and historic work in the most abstract and yet philosophically important area of logic--in a very readable narrative.
This textbook reviews the foundational topics that are typically covered in an introduction to proof course and studies the language of sentential logic as well as investigating the more powerful language of first-order logic and the notion of a formal deduction in first-order logic, in addition, it proves Godel's Completeness Theorem and discusses incompleteness and the computability concept.
Automatic sequences are sequences over a finite alphabet generated by a finite-state machine. This book presents a novel viewpoint on automatic sequences, and more generally on combinatorics on words, by introducing a decision method through which many new results in combinatorics and number theory can be automatically proved or disproved with little or no human intervention. This approach to proving theorems is extremely powerful, allowing long and error-prone case-based arguments to be replaced by simple computations. Readers will learn how to phrase their desired results in first-order logic, using free software to automate the computation process. Results that normally require multipage proofs can emerge in milliseconds, allowing users to engage with mathematical questions that would otherwise be difficult to solve. With more than 150 exercises included, this text is an ideal resource for researchers, graduate students, and advanced undergraduates studying combinatorics, sequences, and number theory.
Using a unique pedagogical approach, this text introduces mathematical logic by guiding students in implementing the underlying logical concepts and mathematical proofs via Python programming. This approach, tailored to the unique intuitions and strengths of the ever-growing population of programming-savvy students, brings mathematical logic into the comfort zone of these students and provides clarity that can only be achieved by a deep hands-on understanding and the satisfaction of having created working code. While the approach is unique, the text follows the same set of topics typically covered in a one-semester undergraduate course, including propositional logic and first-order predicate logic, culminating in a proof of Goedel's completeness theorem. A sneak peek to Goedel's incompleteness theorem is also provided. The textbook is accompanied by an extensive collection of programming tasks, code skeletons, and unit tests. Familiarity with proofs and basic proficiency in Python is assumed.
Floating-point arithmetic is ubiquitous in modern computing, as it is the tool of choice to approximate real numbers. Due to its limited range and precision, its use can become quite involved and potentially lead to numerous failures. One way to greatly increase confidence in floating-point software is by computer-assisted verification of its correctness proofs. This book provides a comprehensive view of how to formally specify and verify tricky floating-point algorithms with the Coq proof assistant. It describes the Flocq formalization of floating-point arithmetic and some methods to automate theorem proofs. It then presents the specification and verification of various algorithms, from error-free transformations to a numerical scheme for a partial differential equation. The examples cover not only mathematical algorithms but also C programs as well as issues related to compilation.
Convergence of Blockchain, AI and IoT: A Digital Platform discusses the convergence of three powerful technologies that play into the digital revolution and blur the lines between biological, digital, and physical objects. This book covers novel algorithms, solutions for addressing issues in applications, security, authentication, and privacy. Discusses innovative technological upgradation and significant challenges in the current era Gives an overview of clinical scientific research that enables smart diagnosis through artificial intelligence Provides an insight into how disruptive technology enabled with the self-running devices and protection mechanism is involved in an augmented reality with blockchain mechanism Talks about neural science being capable of enhancing deep brain waves to predict an overall improvement in human thoughts and behaviours Covers the digital currency mechanism in detail Enhances the knowledge of the readers about smart contract and ledger mechanism with artificial intelligence and blockchain mechanism Targeted audiences range from those interested in the technical revolution of blockchain, big data and the Internet of Things, to research scholars and the professional market.
This is the first of two volumes by Professor Cherlin presenting the state of the art in the classification of homogeneous structures in binary languages and related problems in the intersection of model theory and combinatorics. Researchers and graduate students in the area will find in these volumes many far-reaching results and interesting new research directions to pursue. In this volume, Cherlin develops a complete classification of homogeneous ordered graphs and provides a full proof. He then proposes a new family of metrically homogeneous graphs, a weakening of the usual homogeneity condition. A general classification conjecture is presented, together with general structure theory and applications to a general classification conjecture for such graphs. It also includes introductory chapters giving an overview of the results and methods of both volumes, and an appendix surveying recent developments in the area. An extensive accompanying bibliography of related literature, organized by topic, is available online.
This is the second of two volumes by Professor Cherlin presenting the state of the art in the classification of homogeneous structures in binary languages and related problems in the intersection of model theory and combinatorics. Researchers and graduate students in the area will find in these volumes many far-reaching results and interesting new research directions to pursue. This volume continues the analysis of the first volume to 3-multi-graphs and 3-multi-tournaments, expansions of graphs and tournaments by the addition of a further binary relation. The opening chapter provides an overview of the volume, outlining the relevant results and conjectures. The author applies and extends the results of Volume I to obtain a detailed catalogue of such structures and a second classification conjecture. The book ends with an appendix exploring recent advances and open problems in the theory of homogeneous structures and related subjects.
The International research Library of Philosophy collects in book form a wide range of important and influential essays in philosophy, drawn predominantly from English-language journals. Each volume in the library deals with a field of enquiry which has received significant attention in philosophy in the last 25 years and is edited by a philosopher noted in that field.
Proofs play a central role in advanced mathematics and theoretical computer science, yet many students struggle the first time they take a course in which proofs play a significant role. This bestselling text's third edition helps students transition from solving problems to proving theorems by teaching them the techniques needed to read and write proofs. Featuring over 150 new exercises and a new chapter on number theory, this new edition introduces students to the world of advanced mathematics through the mastery of proofs. The book begins with the basic concepts of logic and set theory to familiarize students with the language of mathematics and how it is interpreted. These concepts are used as the basis for an analysis of techniques that can be used to build up complex proofs step by step, using detailed 'scratch work' sections to expose the machinery of proofs about numbers, sets, relations, and functions. Assuming no background beyond standard high school mathematics, this book will be useful to anyone interested in logic and proofs: computer scientists, philosophers, linguists, and, of course, mathematicians.
Cryptology: Classical and Modern, Second Edition proficiently introduces readers to the fascinating field of cryptology. The book covers classical methods including substitution, transposition, Alberti, Vigenere, and Hill ciphers. It also includes coverage of the Enigma machine, Turing bombe, and Navajo code. Additionally, the book presents modern methods like RSA, ElGamal, and stream ciphers, as well as the Diffie-Hellman key exchange and Advanced Encryption Standard. When possible, the book details methods for breaking both classical and modern methods. The new edition expands upon the material from the first edition which was oriented for students in non-technical fields. At the same time, the second edition supplements this material with new content that serves students in more technical fields as well. Thus, the second edition can be fully utilized by both technical and non-technical students at all levels of study. The authors include a wealth of material for a one-semester cryptology course, and research exercises that can be used for supplemental projects. Hints and answers to selected exercises are found at the end of the book. Features: Requires no prior programming knowledge or background in college-level mathematics Illustrates the importance of cryptology in cultural and historical contexts, including the Enigma machine, Turing bombe, and Navajo code Gives straightforward explanations of the Advanced Encryption Standard, public-key ciphers, and message authentication Describes the implementation and cryptanalysis of classical ciphers, such as substitution, transposition, shift, affine, Alberti, Vigenere, and Hill
A lively and engaging look at logic puzzles and their role in mathematics, philosophy, and recreation Logic puzzles were first introduced to the public by Lewis Carroll in the late nineteenth century and have been popular ever since. Games like Sudoku and Mastermind are fun and engrossing recreational activities, but they also share deep foundations in mathematical logic and are worthy of serious intellectual inquiry. Games for Your Mind explores the history and future of logic puzzles while enabling you to test your skill against a variety of puzzles yourself. In this informative and entertaining book, Jason Rosenhouse begins by introducing readers to logic and logic puzzles and goes on to reveal the rich history of these puzzles. He shows how Carroll's puzzles presented Aristotelian logic as a game for children, yet also informed his scholarly work on logic. He reveals how another pioneer of logic puzzles, Raymond Smullyan, drew on classic puzzles about liars and truthtellers to illustrate Kurt Goedel's theorems and illuminate profound questions in mathematical logic. Rosenhouse then presents a new vision for the future of logic puzzles based on nonclassical logic, which is used today in computer science and automated reasoning to manipulate large and sometimes contradictory sets of data. Featuring a wealth of sample puzzles ranging from simple to extremely challenging, this lively and engaging book brings together many of the most ingenious puzzles ever devised, including the "Hardest Logic Puzzle Ever," metapuzzles, paradoxes, and the logic puzzles in detective stories.
All current methods of secure communication such as public-key cryptography can eventually be broken by faster computing. At the interface of physics and computer science lies a powerful solution for secure communications: quantum cryptography. Because eavesdropping changes the physical nature of the information, users in a quantum exchange can easily detect eavesdroppers. This allows for totally secure random key distribution, a central requirement for use of the one-time pad. Since the one-time pad is theoretically proven to be undecipherable, quantum cryptography is the key to perfect secrecy. Quantum Communications and Cryptography is the first comprehensive review of the past, present, and potential developments in this dynamic field. Leading expert contributors from around the world discuss the scientific foundations, experimental and theoretical developments, and cutting-edge technical and engineering advances in quantum communications and cryptography. The book describes the engineering principles and practical implementations in a real-world metropolitan network as well as physical principles and experimental results of such technologies as entanglement swapping and quantum teleportation. It also offers the first detailed treatment of quantum information processing with continuous variables. Technologies include both free-space and fiber-based communications systems along with the necessary protocols and information processing approaches. Bridging the gap between physics and engineering, Quantum Communications and Cryptography supplies a springboard for further developments and breakthroughs in this rapidly growing area.
Logical paradoxes - like the Liar, Russell's, and the Sorites - are notorious. But in Paradoxes and Inconsistent Mathematics, it is argued that they are only the noisiest of many. Contradictions arise in the everyday, from the smallest points to the widest boundaries. In this book, Zach Weber uses "dialetheic paraconsistency" - a formal framework where some contradictions can be true without absurdity - as the basis for developing this idea rigorously, from mathematical foundations up. In doing so, Weber directly addresses a longstanding open question: how much standard mathematics can paraconsistency capture? The guiding focus is on a more basic question, of why there are paradoxes. Details underscore a simple philosophical claim: that paradoxes are found in the ordinary, and that is what makes them so extraordinary.
Philosophical considerations, which are often ignored or treated casually, are given careful consideration in this introduction. Thomas Forster places the notion of inductively defined sets (recursive datatypes) at the center of his exposition resulting in an original analysis of well established topics. The presentation illustrates difficult points and includes many exercises. Little previous knowledge of logic is required and only a knowledge of standard undergraduate mathematics is assumed.
Between inventing the concept of a universal computer in 1936 and breaking the German Enigma code during World War II, Alan Turing (1912-1954), the British founder of computer science and artificial intelligence, came to Princeton University to study mathematical logic. Some of the greatest logicians in the world--including Alonzo Church, Kurt Godel, John von Neumann, and Stephen Kleene--were at Princeton in the 1930s, and they were working on ideas that would lay the groundwork for what would become known as computer science. This book presents a facsimile of the original typescript of Turing's fascinating and influential 1938 Princeton PhD thesis, one of the key documents in the history of mathematics and computer science. The book also features essays by Andrew Appel and Solomon Feferman that explain the still-unfolding significance of the ideas Turing developed at Princeton. A work of philosophy as well as mathematics, Turing's thesis envisions a practical goal--a logical system to formalize mathematical proofs so they can be checked mechanically. If every step of a theorem could be verified mechanically, the burden on intuition would be limited to the axioms. Turing's point, as Appel writes, is that "mathematical reasoning can be done, and should be done, in mechanizable formal logic." Turing's vision of "constructive systems of logic for practical use" has become reality: in the twenty-first century, automated "formal methods" are now routine. Presented here in its original form, this fascinating thesis is one of the key documents in the history of mathematics and computer science."
In mathematics, we know there are some concepts - objects, constructions, structures, proofs - that are more complex and difficult to describe than others. Computable structure theory quantifies and studies the complexity of mathematical structures, structures such as graphs, groups, and orderings. Written by a contemporary expert in the subject, this is the first full monograph on computable structure theory in 20 years. Aimed at graduate students and researchers in mathematical logic, it brings new results of the author together with many older results that were previously scattered across the literature and presents them all in a coherent framework, making it easier for the reader to learn the main results and techniques in the area for application in their own research. This volume focuses on countable structures whose complexity can be measured within arithmetic; a forthcoming second volume will study structures beyond arithmetic.
Volume II, on formal (ZFC) set theory, incorporates a self-contained "chapter 0" on proof techniques so that it is based on formal logic, in the style of Bourbaki. The emphasis on basic techniques provides a solid foundation in set theory and a thorough context for the presentation of advanced topics (such as absoluteness, relative consistency results, two expositions of Godel's construstive universe, numerous ways of viewing recursion and Cohen forcing).
Constraint Satisfaction Problems (CSPs) are natural computational problems that appear in many areas of theoretical computer science. Exploring which CSPs are solvable in polynomial time and which are NP-hard reveals a surprising link with central questions in universal algebra. This monograph presents a self-contained introduction to the universal-algebraic approach to complexity classification, treating both finite and infinite-domain CSPs. It includes the required background from logic and combinatorics, particularly model theory and Ramsey theory, and explains the recently discovered link between Ramsey theory and topological dynamics and its implications for CSPs. The book will be of interest to graduate students and researchers in theoretical computer science and to mathematicians in logic, combinatorics, and dynamics who wish to learn about the applications of their work in complexity theory.
|
You may like...
Decorative art - Quick. Smart. Fast
Monique Day-Wilde, Angie Franke
Paperback
|