![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Science & Mathematics > Mathematics > Combinatorics & graph theory
The algorithmic solution of problems has always been one of the major concerns of mathematics. For a long time such solutions were based on an intuitive notion of algorithm. It is only in this century that metamathematical problems have led to the intensive search for a precise and sufficiently general formalization of the notions of computability and algorithm. In the 1930s, a number of quite different concepts for this purpose were pro posed, such as Turing machines, WHILE-programs, recursive functions, Markov algorithms, and Thue systems. All these concepts turned out to be equivalent, a fact summarized in Church's thesis, which says that the resulting definitions form an adequate formalization of the intuitive notion of computability. This had and continues to have an enormous effect. First of all, with these notions it has been possible to prove that various problems are algorithmically unsolvable. Among of group these undecidable problems are the halting problem, the word problem theory, the Post correspondence problem, and Hilbert's tenth problem. Secondly, concepts like Turing machines and WHILE-programs had a strong influence on the development of the first computers and programming languages. In the era of digital computers, the question of finding efficient solutions to algorithmically solvable problems has become increasingly important. In addition, the fact that some problems can be solved very efficiently, while others seem to defy all attempts to find an efficient solution, has called for a deeper under standing of the intrinsic computational difficulty of problems."
This book brings together many of the important results in this field. From the reviews: ""A classic gets even better....The edition has new material including the Novelli-Pak-Stoyanovskii bijective proof of the hook formula, Stanley 's proof of the sum of squares formula using differential posets, Fomin 's bijective proof of the sum of squares formula, group acting on posets and their use in proving unimodality, and chromatic symmetric functions." --ZENTRALBLATT MATH
Wick ordering of creation and annihilation operators is of fundamental importance for computing averages and correlations in quantum field theory and, by extension, in the Hudson-Parthasarathy theory of quantum stochastic processes, quantum mechanics, stochastic processes, and probability. This book develops the unified combinatorial framework behind these examples, starting with the simplest mathematically, and working up to the Fock space setting for quantum fields. Emphasizing ideas from combinatorics such as the role of lattice of partitions for multiple stochastic integrals by Wallstrom-Rota and combinatorial species by Joyal, it presents insights coming from quantum probability. It also introduces a 'field calculus' which acts as a succinct alternative to standard Feynman diagrams and formulates quantum field theory (cumulant moments, Dyson-Schwinger equation, tree expansions, 1-particle irreducibility) in this language. Featuring many worked examples, the book is aimed at mathematical physicists, quantum field theorists, and probabilists, including graduate and advanced undergraduate students.
Methods of dimensionality reduction provide a way to understand and visualize the structure of complex data sets. Traditional methods like principal component analysis and classical metric multidimensional scaling suffer from being based on linear models. Until recently, very few methods were able to reduce the data dimensionality in a nonlinear way. However, since the late nineties, many new methods have been developed and nonlinear dimensionality reduction, also called manifold learning, has become a hot topic. New advances that account for this rapid growth are, e.g. the use of graphs to represent the manifold topology, and the use of new metrics like the geodesic distance. In addition, new optimization schemes, based on kernel techniques and spectral decomposition, have lead to spectral embedding, which encompasses many of the recently developed methods. This book describes existing and advanced methods to reduce the dimensionality of numerical databases. For each method, the description starts from intuitive ideas, develops the necessary mathematical details, and ends by outlining the algorithmic implementation. Methods are compared with each other with the help of different illustrative examples. The purpose of the book is to summarize clear facts and ideas about well-known methods as well as recent developments in the topic of nonlinear dimensionality reduction. With this goal in mind, methods are all described from a unifying point of view, in order to highlight their respective strengths and shortcomings. The book is primarily intended for statisticians, computer scientists and data analysts. It is also accessible to other practitioners having a basic background in statistics and/or computational learning, like psychologists (in psychometry) and economists.
Percolation theory is the study of an idealized random medium in two or more dimensions. The emphasis of this book is upon core mathematical material and the presentation of the shortest and most accessible proofs. Much new material appears in this second edition including dynamic and static renormalization, strict inequalities between critical points, a sketch of the lace expansion, and several essays on related fields and applications.
Now fully updated in a third edition, this is a comprehensive textbook on combinatorial optimization. It puts special emphasis on theoretical results and algorithms with provably good performance, in contrast to heuristics. The book contains complete but concise proofs, also for many deep results, some of which have not appeared in print before. Recent topics are covered as well, and numerous references are provided. This third edition contains a new chapter on facility location problems, an area which has been extremely active in the past few years. Furthermore there are several new sections and further material on various topics. New exercises and updates in the bibliography were added.
This title presents new ideas on the visualization of differential equations with user-configurable tools. The authors use the widely-used computer algebra system, Mathematica, to provide an integrated environment for programming, visualizing graphics, and running commentary for learning and working with differential equations.
Natural duality theory is one of the major growth areas within general algebra. This text provides a short path to the forefront of research in duality theory. It presents a coherent approach to new results in the area, as well as exposing open problems. Unary algebras play a special role throughout the text. Individual unary algebras are relatively simple and easy to work with. But as a class they have a rich and complex entanglement with dualisability. This combination of local simplicity and global complexity ensures that, for the study of natural duality theory, unary algebras are an excellent source of examples and counterexamples. A number of results appear here for the first time. In particular, the text ends with an appendix that provides a new and definitive approach to the concept of the rank of a finite algebra and its relationship with strong dualisability.
There has been dramatic growth in the development and application of Bayesian inference in statistics. Berger (2000) documents the increase in Bayesian activity by the number of published research articles, the number of books, andtheextensivenumberofapplicationsofBayesianarticlesinapplied disciplines such as science and engineering. One reason for the dramatic growth in Bayesian modeling is the availab- ity of computational algorithms to compute the range of integrals that are necessary in a Bayesian posterior analysis. Due to the speed of modern c- puters, it is now possible to use the Bayesian paradigm to 't very complex models that cannot be 't by alternative frequentist methods. To 't Bayesian models, one needs a statistical computing environment. This environment should be such that one can: write short scripts to de?ne a Bayesian model use or write functions to summarize a posterior distribution use functions to simulate from the posterior distribution construct graphs to illustrate the posterior inference An environment that meets these requirements is the R system. R provides a wide range of functions for data manipulation, calculation, and graphical d- plays. Moreover, it includes a well-developed, simple programming language that users can extend by adding new functions. Many such extensions of the language in the form of packages are easily downloadable from the Comp- hensive R Archive Network (CRAN)
Leave nothing to chance. This cliche embodies the common belief that ran domness has no place in carefully planned methodologies, every step should be spelled out, each i dotted and each t crossed. In discrete mathematics at least, nothing could be further from the truth. Introducing random choices into algorithms can improve their performance. The application of proba bilistic tools has led to the resolution of combinatorial problems which had resisted attack for decades. The chapters in this volume explore and celebrate this fact. Our intention was to bring together, for the first time, accessible discus sions of the disparate ways in which probabilistic ideas are enriching discrete mathematics. These discussions are aimed at mathematicians with a good combinatorial background but require only a passing acquaintance with the basic definitions in probability (e.g. expected value, conditional probability). A reader who already has a firm grasp on the area will be interested in the original research, novel syntheses, and discussions of ongoing developments scattered throughout the book. Some of the most convincing demonstrations of the power of these tech niques are randomized algorithms for estimating quantities which are hard to compute exactly. One example is the randomized algorithm of Dyer, Frieze and Kannan for estimating the volume of a polyhedron. To illustrate these techniques, we consider a simple related problem. Suppose S is some region of the unit square defined by a system of polynomial inequalities: Pi (x. y) ~ o.
From the reviews: " A unique and fascinating blend, which is shown to be useful for a variety of applications, including robotics, geometrical optics, computer animation, and geometric design. The contents of the book are visualized by a wealth of carefully chosen illustrations, making the book a shear pleasure to read, or even to just browse in." Mathematical Reviews
Image analysis is a computational feat which humans show excellence in, in comp- ison with computers. Yet the list of applications that rely on automatic processing of images has been growing at a fast pace. Biometric authentication by face, ?ngerprint, and iris, online character recognition in cell phones as well as drug design tools are but a few of its benefactors appearing on the headlines. This is, of course, facilitated by the valuable output of the resarch community in the past 30 years. The pattern recognition and computer vision communities that study image analysis have large conferences, which regularly draw 1000 parti- pants. In a way this is not surprising, because much of the human-speci?c activities critically rely on intelligent use of vision. If routine parts of these activities can be automated, much is to be gained in comfort and sustainable development. The - search ?eld could equally be called visualintelligence because it concerns nearly all activities of awake humans. Humans use or rely on pictures or pictorial languages to represent, analyze, and develop abstract metaphors related to nearly every aspect of thinking and behaving, be it science, mathematics, philosopy, religion, music, or emotions. The present volume is an introductory textbook on signal analysis of visual c- putation for senior-level undergraduates or for graduate students in science and - gineering. My modest goal has been to present the frequently used techniques to analyze images in a common framework-directional image processing.
What is the "most uniform" way of distributing n points in the unit square? How big is the "irregularity" necessarily present in any such distribution? This book is an accessible and lively introduction to the area of geometric discrepancy theory, with numerous exercises and illustrations. In separate, more specialized parts, it also provides a comprehensive guide to recent research.
The lectures concentrate on highlights in Combinatorial (ChaptersII and III) and Number Theoretical (ChapterIV) Extremal Theory, in particular on the solution of famous problems which were open for many decades. However, the organization of the lectures in six chapters does neither follow the historic developments nor the connections between ideas in several cases. With the speci?ed auxiliary results in ChapterI on Probability Theory, Graph Theory, etc., all chapters can be read and taught independently of one another. In addition to the 16 lectures organized in 6 chapters of the main part of the book, there is supplementary material for most of them in the Appendix. In parti- lar, there are applications and further exercises, research problems, conjectures, and even research programs. The following books and reports [B97], [ACDKPSWZ00], [A01], and [ABCABDM06], mostly of the authors, are frequently cited in this book, especially in the Appendix, and we therefore mark them by short labels as [B], [N], [E], and [G]. We emphasize that there are also "Exercises" in [B], a "Problem Section" with contributions by several authors on pages 1063-1105 of [G], which are often of a combinatorial nature, and "Problems and Conjectures" on pages 172-173 of [E].
A matroid is an abstract mathematical structure that captures combinatorial properties of matrices. This book offers a unique introduction to matroid theory, emphasizing motivations from matrix theory and applications to systems analysis. This book serves also as a comprehensive presentation of the theory and application of mixed matrices, developed primarily by the present author in the 1990's. A mixed matrix is a convenient mathematical tool for systems analysis, compatible with the physical observation that "fixed constants" and "system parameters" are to be distinguished in the description of engineering systems. This book will be extremely useful to graduate students and researchers in engineering, mathematics and computer science. From the reviews: ..".The book has been prepared very carefully, contains a lot of interesting results and is highly recommended for graduate and postgraduate students." Andras Recski, Mathematical Reviews Clippings 2000m:93006"
This book illustrates how modern mathematical wavelet transform techniques offer fresh insights into the complex behavior of neural systems at different levels: from the microscopic dynamics of individual cells to the macroscopic behavior of large neural networks. It also demonstrates how and where wavelet-based mathematical tools can provide an advantage over classical approaches used in neuroscience. The authors well describe single neuron and populational neural recordings. This 2nd edition discusses novel areas and significant advances resulting from experimental techniques and computational approaches developed since 2015, and includes three new topics: * Detection of fEPSPs in multielectrode LFPs recordings. * Analysis of Visual Sensory Processing in the Brain and BCI for Human Attention Control; * Analysis and Real-time Classification of Motor-related EEG Patterns; The book is a valuable resource for neurophysiologists and physicists familiar with nonlinear dynamical systems and data processing, as well as for graduate students specializing in these and related areas.
This volume surveys the development of combinatorics since 1930 by presenting in chronological order the fundamental results of the subject proved in over five decades of original papers by: T. van Aardenne-Ehrenfest.- R.L. Brooks.- N.G. de Bruijn.- G.F. Clements.- H.H. Crapo.- R.P. Dilworth.- J. Edmonds.- P. Erd s.- L.R. Ford, Jr.- D.R. Fulkerson.- D. Gale.- L. Geissinger.- I.J. Good.- R.L. Graham.- A.W. Hales.- P. Hall.- P.R. Halmos.- R.I. Jewett.- I. Kaplansky.- P.W. Kasteleyn.- G. Katona.- D.J. Kleitman.- K. Leeb.- B. Lindstr m.- L. Lov sz.- D. Lubell.- C. St. J.A. Nash-Williams.- G. P lya.-R. Rado.- F.P. Ramsey.- G.-C. Rota.- B.L. Rothschild.- H.J. Ryser.- C. Schensted.- M.P. Sch tzenberger.- R.P. Stanley.- G. Szekeres.- W.T. Tutte.- H.E. Vaughan.- H. Whitney.
A brilliant treatment of a knotty problem in computing. This volume contains chapters written by reputable researchers and provides the state of the art in theory and algorithms for the traveling salesman problem (TSP). The book covers all important areas of study on TSP, including polyhedral theory for symmetric and asymmetric TSP, branch and bound, and branch and cut algorithms, probabilistic aspects of TSP, and includes a thorough computational analysis of heuristic and metaheuristic algorithms.
Marking 94 years since its first appearance, this book provides an annotated translation of Sainte-Lague's seminal monograph Les reseaux (ou graphes), drawing attention to its fundamental principles and ideas. Sainte-Lague's 1926 monograph appeared only in French, but in the 1990s H. Gropp published a number of English papers describing several aspects of the book. He expressed his hope that an English translation might sometime be available to the mathematics community. In the 10 years following the appearance of Les reseaux (ou graphes), the development of graph theory continued, culminating in the publication of the first full book on the theory of finite and infinite graphs in 1936 by Denes Koenig. This remained the only well-known text until Claude Berge's 1958 book on the theory and applications of graphs. By 1960, graph theory had emerged as a significant mathematical discipline of its own. This book will be of interest to graph theorists and mathematical historians.
Outer billiards provides a toy model for planetary motion and exhibits intricate and mysterious behavior even for seemingly simple examples. It is a dynamical system in which a particle in the plane moves around the outside of a convex shape according to a scheme that is reminiscent of ordinary billiards. The Plaid Model, which is a self-contained sequel to Richard Schwartz's Outer Billiards on Kites, provides a combinatorial model for orbits of outer billiards on kites. Schwartz relates these orbits to such topics as polytope exchange transformations, renormalization, continued fractions, corner percolation, and the Truchet tile system. The combinatorial model, called "the plaid model," has a self-similar structure that blends geometry and elementary number theory. The results were discovered through computer experimentation and it seems that the conclusions would be extremely difficult to reach through traditional mathematics. The book includes an extensive computer program that allows readers to explore the materials interactively and each theorem is accompanied by a computer demonstration.
How many possible sudoku puzzles are there? In the lottery, what is the chance that two winning balls have consecutive numbers? Who invented Pascal's triangle? (it was not Pascal) Combinatorics, the branch of mathematics concerned with selecting, arranging, and listing or counting collections of objects, works to answer all these questions. Dating back some 3000 years, and initially consisting mainly of the study of permutations and combinations, its scope has broadened to include topics such as graph theory, partitions of numbers, block designs, design of codes, and latin squares. In this Very Short Introduction Robin Wilson gives an overview of the field and its applications in mathematics and computer theory, considering problems from the shortest routes covering certain stops to the minimum number of colours needed to colour a map with different colours for neighbouring countries. ABOUT THE SERIES: The Very Short Introductions series from Oxford University Press contains hundreds of titles in almost every subject area. These pocket-sized books are the perfect way to get ahead in a new subject quickly. Our expert authors combine facts, analysis, perspective, new ideas, and enthusiasm to make interesting and challenging topics highly readable.
This is the first book that focuses entirely on the fundamental questions in visualization. Unlike other existing books in the field, it contains discussions that go far beyond individual visual representations and individual visualization algorithms. It offers a collection of investigative discourses that probe these questions from different perspectives, including concepts that help frame these questions and their potential answers, mathematical methods that underpin the scientific reasoning of these questions, empirical methods that facilitate the validation and falsification of potential answers, and case studies that stimulate hypotheses about potential answers while providing practical evidence for such hypotheses. Readers are not instructed to follow a specific theory, but their attention is brought to a broad range of schools of thoughts and different ways of investigating fundamental questions. As such, the book represents the by now most significant collective effort for gathering a large collection of discourses on the foundation of data visualization. Data visualization is a relatively young scientific discipline. Over the last three decades, a large collection of computer-supported visualization techniques have been developed, and the merits and benefits of using these techniques have been evidenced by numerous applications in practice. These technical advancements have given rise to the scientific curiosity about some fundamental questions such as why and how visualization works, when it is useful or effective and when it is not, what are the primary factors affecting its usefulness and effectiveness, and so on. This book signifies timely and exciting opportunities to answer such fundamental questions by building on the wealth of knowledge and experience accumulated in developing and deploying visualization technology in practice.
A Path to Combinatorics for Undergraduates is a lively introduction not only to combinatorics, but also to mathematical ingenuity, rigor, and the joy of solving puzzles. This unique approach to combinatorics is centered around unconventional, essay-type combinatorial examples, followed by a number of carefully selected, challenging problems and extensive discussions of their solutions. Topics encompass permutations and combinations, binomial coefficients and their applications, bijections, inclusions and exclusions, and generating functions. Each chapter features fully-worked problems, including many from Olympiads and other competitions, as well as a number of problems original to the authors; at the end of each chapter are further exercises to reinforce understanding, encourage creativity, and build a repertory of problem-solving techniques. The authors' previous text, 102 Combinatorial Problems, makes a fine companion volume to the present work, which is ideal for Olympiad participants and coaches, advanced high school students, undergraduates, and college instructors. The book's unusual problems and examples will interest seasoned mathematicians as well.
This work offers a contribution in the geometric form of the theory of several complex variables. Since complex Grassmann manifolds serve as classifying spaces of complex vector bundles, the cohomology structure of a complex Grassmann manifold is of importance for the construction of Chern classes of complex vector bundles. The cohomology ring of a Grassmannian is therefore of interest in topology, differential geometry, algebraic geometry, and complex analysis. Wilhelm Stoll treats certain aspects of the complex analysis point of view. This work originated with questions in value distribution theory. Here analytic sets and differential forms rather than the corresponding homology and cohomology classes are considered. On the Grassmann manifold, the cohomology ring is isomorphic to the ring of differential forms invariant under the unitary group, and each cohomology class is determined by a family of analytic sets.
"102 Combinatorial Problems" consists of carefully selected problems that have been used in the training and testing of the USA International Mathematical Olympiad (IMO) team. Key features: * Provides in-depth enrichment in the important areas of combinatorics by reorganizing and enhancing problem-solving tactics and strategies * Topics include: combinatorial arguments and identities, generating functions, graph theory, recursive relations, sums and products, probability, number theory, polynomials, theory of equations, complex numbers in geometry, algorithmic proofs, combinatorial and advanced geometry, functional equations and classical inequalities The book is systematically organized, gradually building combinatorial skills and techniques and broadening the student's view of mathematics. Aside from its practical use in training teachers and students engaged in mathematical competitions, it is a source of enrichment that is bound to stimulate interest in a variety of mathematical areas that are tangential to combinatorics. |
You may like...
Fixed Point Theory and Graph Theory…
Monther Alfuraidan, Qamrul Ansari
Hardcover
R1,860
Discovery Miles 18 600
|