![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Applications of computing > Artificial intelligence > Neural networks
Much research focuses on the question of how information is processed in nervous systems, from the level of individual ionic channels to large-scale neuronal networks, and from "simple" animals such as sea slugs and flies to cats and primates. New interdisciplinary methodologies combine a bottom-up experimental methodology with the more top-down-driven computational and modeling approach. This book serves as a handbook of computational methods and techniques for modeling the functional properties of single and groups of nerve cells.The contributors highlight several key trends: (1) the tightening link between analytical/numerical models and the associated experimental data, (2) the broadening of modeling methods, at both the subcellular level and the level of large neuronal networks that incorporate real biophysical properties of neurons as well as the statistical properties of spike trains, and (3) the organization of the data gained by physical emulation of the nervous system components through the use of very large scale circuit integration (VLSI) technology.The field of neuroscience has grown dramatically since the first edition of this book was published nine years ago. Half of the chapters of the second edition are completely new; the remaining ones have all been thoroughly revised. Many chapters provide an opportunity for interactive tutorials and simulation programs. They can be accessed via Christof Koch's Website.Contributors: Larry F. Abbott, Paul R. Adams, Hagai Agmon-Snir, James M. Bower, Robert E. Burke, Erik de Schutter, Alain Destexhe, Rodney Douglas, Bard Ermentrout, Fabrizio Gabbiani, David Hansel, Michael Hines, Christof Koch, Misha Mahowald, Zachary F. Mainen, Eve Marder, Michael V. Mascagni, Alexander D. Protopapas, Wilfrid Rall, John Rinzel, Idan Segev, Terrence J. Sejnowski, Shihab Shamma, Arthur S. Sherman, Paul Smolen, Haim Sompolinsky, Michael Vanier, Walter M. Yamada.
This textbook provides a thorough introduction to the field of learning from experimental data and soft computing. Support vector machines (SVM) and neural networks (NN) are the mathematical structures, or models, that underlie learning, while fuzzy logic systems (FLS) enable us to embed structured human knowledge into workable algorithms. The book assumes that it is not only useful, but necessary, to treat SVM, NN, and FLS as parts of a connected whole. Throughout, the theory and algorithms are illustrated by practical examples, as well as by problem sets and simulated experiments. This approach enables the reader to develop SVM, NN, and FLS in addition to understanding them. The book also presents three case studies: on NN-based control, financial time series analysis, and computer graphics. A solutions manual and all of the MATLAB programs needed for the simulated experiments are available.
Most practical applications of artificial neural networks are based on a computational model involving the propagation of continuous variables from one processing unit to the next. In recent years, data from neurobiological experiments have made it increasingly clear that biological neural networks, which communicate through pulses, use the timing of the pulses to transmit information and perform computation. This realization has stimulated significant research on pulsed neural networks, including theoretical analyses and model development, neurobiological modeling, and hardware implementation. This book presents the complete spectrum of current research in pulsed neural networks and includes the most important work from many of the key scientists in the field. Terrence J. Sejnowski's foreword, "Neural Pulse Coding," presents an overview of the topic. The first half of the book consists of longer tutorial articles spanning neurobiology, theory, algorithms, and hardware. The second half contains a larger number of shorter research chapters that present more advanced concepts. The contributors use consistent notation and terminology throughout the book. Contributors Peter S. Burge, Stephen R. Deiss, Rodney J. Douglas, John G. Elias, Wulfram Gerstner, Alister Hamilton, David Horn, Axel Jahnke, Richard Kempter, Wolfgang Maass, Alessandro Mortara, Alan F. Murray, David P. M. Northmore, Irit Opher, Kostas A. Papathanasiou, Michael Recce, Barry J. P. Rising, Ulrich Roth, Tim Schoenauer, Terrence J. Sejnowski, John Shawe-Taylor, Max R. van Daalen, J. Leo van Hemmen, Philippe Venier, Hermann Wagner, Adrian M. Whatley, Anthony M. Zador
The first truly up-to-date look at the theory and capabilities of nonlinear dynamical systems that take the form of feedforward neural network structures Considered one of the most important types of structures in the study of neural networks and neural-like networks, feedforward networks incorporating dynamical elements have important properties and are of use in many applications. Specializing in experiential knowledge, a neural network stores and expands its knowledge base via strikingly human routes–through a learning process and information storage involving interconnection strengths known as synaptic weights. In Nonlinear Dynamical Systems: Feedforward Neural Network Perspectives, six leading authorities describe recent contributions to the development of an analytical basis for the understanding and use of nonlinear dynamical systems of the feedforward type, especially in the areas of control, signal processing, and time series analysis. Moving from an introductory discussion of the different aspects of feedforward neural networks, the book then addresses:
It then sheds light on the application of feedforward neural networks to speech processing, summarizing speech-related techniques, and reviewing feedforward neural networks from the viewpoint of fundamental design issues. An up-to-date and authoritative look at the ever-widening technical boundaries and influence of neural networks in dynamical systems, this volume is an indispensable resource for researchers in neural networks and a reference staple for libraries.
Surprising tales from the scientists who first learned how to use computers to understand the workings of the human brain.Since World War II, a group of scientists has been attempting to understand the human nervous system and to build computer systems that emulate the brain's abilities. Many of the early workers in this field of neural networks came from cybernetics; others came from neuroscience, physics, electrical engineering, mathematics, psychology, even economics. In this collection of interviews, those who helped to shape the field share their childhood memories, their influences, how they became interested in neural networks, and what they see as its future.The subjects tell stories that have been told, referred to, whispered about, and imagined throughout the history of the field. Together, the interviews form a Rashomon-like web of reality. Some of the mythic people responsible for the foundations of modern brain theory and cybernetics, such as Norbert Wiener, Warren McCulloch, and Frank Rosenblatt, appear prominently in the recollections. The interviewees agree about some things and disagree about more. Together, they tell the story of how science is actually done, including the false starts, and the Darwinian struggle for jobs, resources, and reputation. Although some of the interviews contain technical material, there is no actual mathematics in the book.ContributorsJames A. Anderson, Michael Arbib, Gail Carpenter, Leon Cooper, Jack Cowan, Walter Freeman, Stephen Grossberg, Robert Hecht-Neilsen, Geoffrey Hinton, Teuvo Kohonen, Bart Kosko, Jerome Lettvin, Carver Mead, David Rumelhart, Terry Sejnowski, Paul Werbos, Bernard Widrow
a short and accessible introduction on AI and Cars written by leading experts
Since its founding in 1989 by Terrence Sejnowski, Neural Computation has become the leading journal in the field. Foundations of Neural Computationcollects, by topic, the most significant papers that have appeared in the journal over the past nine years.This volume of Foundations of Neural Computation, on unsupervised learning algorithms, focuses on neural network learning algorithms that do not require an explicit teacher. The goal of unsupervised learning is to extract an efficient internal representation of the statistical structure implicit in the inputs. These algorithms provide insights into the development of the cerebral cortex and implicit learning in humans. They are also of interest to engineers working in areas such as computer vision and speech recognition who seek efficient representations of raw input data.
This book provides a comprehensive introduction to the computational material that forms the underpinnings of the currently evolving set of brain models. It is now clear that the brain is unlikely to be understood without recourse to computational theories. The theme of An Introduction to Natural Computation is that ideas from diverse areas such as neuroscience, information theory, and optimization theory have recently been extended in ways that make them useful for describing the brains programs. This book provides a comprehensive introduction to the computational material that forms the underpinnings of the currently evolving set of brain models. It stresses the broad spectrum of learning models-ranging from neural network learning through reinforcement learning to genetic learning-and situates the various models in their appropriate neural context. To write about models of the brain before the brain is fully understood is a delicate matter. Very detailed models of the neural circuitry risk losing track of the task the brain is trying to solve. At the other extreme, models that represent cognitive constructs can be so abstract that they lose all relationship to neurobiology. An Introduction to Natural Computation takes the middle ground and stresses the computational task while staying near the neurobiology.
Optimization Techniques is a unique reference source to a diverse
array of methods for achieving optimization, and includes both
systems structures and computational methods. The text devotes
broad coverage toa unified view of optimal learning, orthogonal
transformation techniques, sequential constructive techniques, fast
back propagation algorithms, techniques for neural networks with
nonstationary or dynamic outputs, applications to constraint
satisfaction, optimization issues and techniques for unsupervised
learning neural networks, optimum Cerebellar Model of Articulation
Controller systems, a new statistical theory of optimum neural
learning, and the role of the Radial Basis Function in nonlinear
dynamical systems.This volume is useful for practitioners,
researchers, and students in industrial, manufacturing, mechanical,
electrical, and computer engineering.
Image Processing and Pattern Recognition covers major applications
in the field, including optical character recognition, speech
classification, medical imaging, paper currency recognition,
classification reliability techniques, and sensor technology. The
text emphasizes algorithms and architectures for achieving
practical and effective systems, and presents many examples.
Practitioners, researchers, and students in computer science,
electrical engineering, andradiology, as well as those working at
financial institutions, will value this unique and authoritative
reference to diverse applications methodologies.
Industrial and Manufacturing Systems serves as an in-depth guide to
major applications in this focal area of interest to the
engineering community. This volume emphasizes the neural network
structures used to achieve practical and effective systems, and
provides numerous examples. Industrial and Manufacturing Systems is
a unique and comprehensive reference to diverse application
methodologies and implementations by means of neural network
systems. It willbe of use to practitioners, researchers, and
students in industrial, manufacturing, electrical, and mechanical
engineering, as well as in computer science and engineering.
The book emphasizes neural network structures for achieving
practical and effective systems, and provides many examples.
Practitioners, researchers, and students in industrial,
manufacturing, electrical, mechanical, and production engineering
will find this volume a unique and comprehensive reference source
for diverse application methodologies.
This volume links the concept of granular computing using deep learning and the Internet of Things to object tracking for video analysis. It describes how uncertainties, involved in the task of video processing, could be handled in rough set theoretic granular computing frameworks. Issues such as object tracking from videos in constrained situations, occlusion/overlapping handling, measuring of the reliability of tracking methods, object recognition and linguistic interpretation in video scenes, and event prediction from videos, are the addressed in this volume. The book also looks at ways to reduce data dependency in the context of unsupervised (without manual interaction/ labeled data/ prior information) training.This book may be used both as a textbook and reference book for graduate students and researchers in computer science, electrical engineering, system science, data science, and information technology, and is recommended for both students and practitioners working in computer vision, machine learning, video analytics, image analytics, artificial intelligence, system design, rough set theory, granular computing, and soft computing.
This book is one of the most up-to-date and cutting-edge texts
available on the rapidly growing application area of neural
networks. Neural Networks and Pattern Recognition focuses on the
use of neural networksin pattern recognition, a very important
application area for neural networks technology. The contributors
are widely known and highly respected researchers and practitioners
in the field.
Neural networks are an exciting technology of growing importance in real industrial situations, particularly in control and systems. This book aims to give a detailed appreciation of the use of neural nets in these applications; it is aimed particularly at those with a control or systems background who wish to gain an insight into the technology in the context of real applications. The book introduces a wide variety of network types, including Kohenen nets, n-tuple nets and radial basis function networks, as well as the more usual multi-layer perception back-propagation networks. It begins by describing the basic principles and some essential design features, then goes on to examine in depth several application studies illustrating a range of advanced approaches to the topic.
Neural Systems for Robotics represents the most up-to-date
developments in the rapidly growing aplication area of neural
networks, which is one of the hottest application areas for neural
networks technology. The book not only contains a comprehensive
study of neurocontrollers in complex Robotics systems, written by
highly respected researchers in the field but outlines a novel
approach to solving Robotics problems. The importance of neural
networks in all aspects of Robot arm manipulators, neurocontrol,
and Robotic systems is also given thorough and in-depth coverage.
All researchers and students dealing with Robotics will find Neural
Systems for Robotics of immense interest and assistance.
Control problems offer an industrially important application and a
guide to understanding control systems for those working in Neural
Networks. Neural Systems for Control represents the most up-to-date
developments in the rapidly growing aplication area of neural
networks and focuses on research in natural and artifical neural
systems directly applicable to control or making use of modern
control theory. The book covers such important new developments in
control systems such as intelligent sensors in semiconductor wafer
manufacturing; the relation between muscles and cerebral neurons in
speech recognition; online compensation of reconfigurable control
for spacecraft aircraft and other systems; applications to rolling
mills, robotics and process control; the usage of past output data
to identify nonlinear systems by neural networks; neural
approximate optimal control; model-free nonlinear control; and
neural control based on a regulation of physiological
investigation/blood pressure control. All researchers and students
dealing with control systems will find the fascinating Neural
Systems for Control of immense interest and assistance.
This book is the companion volume to "Rethinking Innateness: A Connectionist Perspective on Development" (The MIT Press, 1996), which proposed a new theoretical framework to answer the question "What does it mean to say that a behavior is innate?" The new work provides concrete illustrations -- in the form of computer simulations -- of properties of connectionist models that are particularly relevant to cognitive development. This enables the reader to pursue in depth some of the practical and empirical issues raised in the first book. The authors' larger goal is to demonstrate the usefulness of neural network modeling as a research methodology. The book comes with a complete software package, including demonstration projects, for running neural network simulations on both Macintosh and Windows 95. It also contains a series of exercises in the use of the neural network simulator provided with the book. The software is also available to run on a variety of UNIX platforms.
Stephen Grossberg and his colleagues at Boston University's Center for Adaptive Systems are producing some of the most exciting research in the neural network approach to making computers "think." Packed with real-time computer simulations and rigorous demonstrations of these phenomena, this book includes results on vision, speech, cognitive information processing; adaptive pattern recognition, adaptive robotics, conditioning and attention, cognitive-emotional interactions, and decision making under risk.
Theoretical Mechanics of Biological Neural Networks develops an engineering science for the description of neuroclectric signalling of biological neural networks in terms of the underlying neurobiological mechanisms. The primary theoretical contribution of the book is to show how to describe the co-ordinated electrical activity of arbitrarily complex neural networks in terms of a single governing principle ' for each significant component in the same way that Newton's formulation of classical mechanics allows one to express force-motion relationships for arbitrarily complex mechanical systems in terms of one fundamental principle of motion for each constituent element.;Practically, the book shows how to generate mathematical and computational representations Of' the co-ordinated electrical activity of neural networks, ranging from individual neurons to composite systems of interconnected networks. Complete listings of several general purpose computer programs embodying the theory are included in the book.
This unique compendium represents important action of fuzzy systems to quantum mechanics. From fuzzy sets to fuzzy systems, it also gives clear descriptions on the development on fuzzy logic, where the most important result is the probability presentation of fuzzy systems.The important conclusions on fuzzy systems are used in the study of quantum mechanics, which is a very new idea. Eight important conclusions are obtained. The author has proved that mass-point motions in classical mechanics must have waves, which means that any mass-point motion in classical mechanics has wave mass-point dualism as well as any microscopic particle motion must have wave-particle dualism. Based on this conclusion, it has been proven that classical mechanics and quantum mechanics are unified.
This current book provides new research on artificial neural networks (ANNs). Topics discussed include the application of ANNs in chemistry and chemical engineering fields; the application of ANNs in the prediction of biodiesel fuel properties from fatty acid constituents; the use of ANNs for solar radiation estimation; the use of in silico methods to design and evaluate skin UV filters; a practical model based on the multilayer perceptron neural network (MLP) approach to predict the milling tool flank wear in a regular cut, as well as entry cut and exit cut, of a milling tool; parameter extraction of small-signal and noise models of microwave transistors based on ANNs; and the application of ANNs to deep-learning and predictive analysis in semantic TCM telemedicine systems. |
![]() ![]() You may like...
Handbook of Research on Advanced…
Madhumangal Pal, Sovan Samanta, …
Hardcover
R7,278
Discovery Miles 72 780
Intelligent Analysis Of Fundus Images…
Yuanyuan Chen, Yi Zhang, …
Hardcover
R2,321
Discovery Miles 23 210
Fuzzy Knowledge Management for the…
Zongmin Ma, Fu Zhang, …
Hardcover
Fuzzy Systems - Theory and Applications
Constantin Volosencu
Hardcover
R3,370
Discovery Miles 33 700
Icle Publications Plc-Powered Data…
Polly Patrick, Angela Peery
Paperback
R758
Discovery Miles 7 580
|