|
Showing 1 - 9 of
9 matches in All Departments
This book brings together historical notes, reviews of research
developments, fresh ideas on how to make VC (Vapnik-Chervonenkis)
guarantees tighter, and new technical contributions in the areas of
machine learning, statistical inference, classification,
algorithmic statistics, and pattern recognition. The contributors
are leading scientists in domains such as statistics, mathematics,
and theoretical computer science, and the book will be of interest
to researchers and graduate students in these domains.
This book honours the outstanding contributions of Vladimir Vapnik,
a rare example of a scientist for whom the following statements
hold true simultaneously: his work led to the inception of a new
field of research, the theory of statistical learning and empirical
inference; he has lived to see the field blossom; and he is still
as active as ever. He started analyzing learning algorithms in the
1960s and he invented the first version of the generalized portrait
algorithm. He later developed one of the most successful methods in
machine learning, the support vector machine (SVM) - more than just
an algorithm, this was a new approach to learning problems,
pioneering the use of functional analysis and convex optimization
in machine learning. Part I of this book contains three chapters
describing and witnessing some of Vladimir Vapnik's contributions
to science. In the first chapter, Leon Bottou discusses the seminal
paper published in 1968 by Vapnik and Chervonenkis that lay the
foundations of statistical learning theory, and the second chapter
is an English-language translation of that original paper. In the
third chapter, Alexey Chervonenkis presents a first-hand account of
the early history of SVMs and valuable insights into the first
steps in the development of the SVM in the framework of the
generalised portrait method. The remaining chapters, by leading
scientists in domains such as statistics, theoretical computer
science, and mathematics, address substantial topics in the theory
and practice of statistical learning theory, including SVMs and
other kernel-based methods, boosting, PAC-Bayesian theory, online
and transductive learning, loss functions, learnable function
classes, notions of complexity for function classes, multitask
learning, and hypothesis selection. These contributions include
historical and context notes, short surveys, and comments on future
research directions. This book will be of interest to researchers,
engineers, and graduate students engaged with all aspects of
statistical learning.
Algorithmic Learning in a Random World describes recent theoretical
and experimental developments in building computable approximations
to Kolmogorov's algorithmic notion of randomness. Based on these
approximations, a new set of machine learning algorithms have been
developed that can be used to make predictions and to estimate
their confidence and credibility in high-dimensional spaces under
the usual assumption that the data are independent and identically
distributed (assumption of randomness). Another aim of this unique
monograph is to outline some limits of predictions: The approach
based on algorithmic theory of randomness allows for the proof of
impossibility of prediction in certain situations. The book
describes how several important machine learning problems, such as
density estimation in high-dimensional spaces, cannot be solved if
the only assumption is randomness.
This book is about conformal prediction, an approach to prediction
that originated in machine learning in the late 1990s. The main
feature of conformal prediction is the principled treatment of the
reliability of predictions. The prediction algorithms described -
conformal predictors - are provably valid in the sense that they
evaluate the reliability of their own predictions in a way that is
neither over-pessimistic nor over-optimistic (the latter being
especially dangerous). The approach is still flexible enough to
incorporate most of the existing powerful methods of machine
learning. The book covers both key conformal predictors and the
mathematical analysis of their properties. Algorithmic Learning in
a Random World contains, in addition to proofs of validity, results
about the efficiency of conformal predictors. The only assumption
required for validity is that of "randomness" (the prediction
algorithm is presented with independent and identically distributed
examples); in later chapters, even the assumption of randomness is
significantly relaxed. Interesting results about efficiency are
established both under randomness and under stronger assumptions.
Since publication of the First Edition in 2005 conformal prediction
has found numerous applications in medicine and industry, and is
becoming a popular machine-learning technique. This Second Edition
contains three new chapters. One is about conformal predictive
distributions, which are more informative than the set predictions
produced by standard conformal predictors. Another is about the
efficiency of ways of testing the assumption of randomness based on
conformal prediction. The third new chapter harnesses conformal
testing procedures for protecting machine-learning algorithms
against changes in the distribution of the data. In addition, the
existing chapters have been revised, updated, and expanded.
The conformal predictions framework is a recent development in
machine learning that can associate a reliable measure of
confidence with a prediction in any real-world pattern recognition
application, including risk-sensitive applications such as medical
diagnosis, face recognition, and financial risk prediction.
"Conformal Predictions for Reliable Machine Learning: Theory,
Adaptations and Applications" captures the basic theory of the
framework, demonstrates how to apply it to real-world problems, and
presents several adaptations, including active learning, change
detection, and anomaly detection. As practitioners and researchers
around the world apply and adapt the framework, this edited volume
brings together these bodies of work, providing a springboard for
further research as well as a handbook for application in
real-world problems.
Understand the theoretical foundations of this important framework
that can provide a reliable measure of confidence with predictions
in machine learningBe able to apply this framework to real-world
problems in different machine learning settings, including
classification, regression, and clusteringLearn effective ways of
adapting the framework to newer problem settings, such as active
learning, model selection, or change detection
This book brings together historical notes, reviews of research
developments, fresh ideas on how to make VC (Vapnik-Chervonenkis)
guarantees tighter, and new technical contributions in the areas of
machine learning, statistical inference, classification,
algorithmic statistics, and pattern recognition. The contributors
are leading scientists in domains such as statistics, mathematics,
and theoretical computer science, and the book will be of interest
to researchers and graduate students in these domains.
This book honours the outstanding contributions of Vladimir Vapnik,
a rare example of a scientist for whom the following statements
hold true simultaneously: his work led to the inception of a new
field of research, the theory of statistical learning and empirical
inference; he has lived to see the field blossom; and he is still
as active as ever. He started analyzing learning algorithms in the
1960s and he invented the first version of the generalized portrait
algorithm. He later developed one of the most successful methods in
machine learning, the support vector machine (SVM) - more than just
an algorithm, this was a new approach to learning problems,
pioneering the use of functional analysis and convex optimization
in machine learning. Part I of this book contains three chapters
describing and witnessing some of Vladimir Vapnik's contributions
to science. In the first chapter, Leon Bottou discusses the seminal
paper published in 1968 by Vapnik and Chervonenkis that lay the
foundations of statistical learning theory, and the second chapter
is an English-language translation of that original paper. In the
third chapter, Alexey Chervonenkis presents a first-hand account of
the early history of SVMs and valuable insights into the first
steps in the development of the SVM in the framework of the
generalised portrait method. The remaining chapters, by leading
scientists in domains such as statistics, theoretical computer
science, and mathematics, address substantial topics in the theory
and practice of statistical learning theory, including SVMs and
other kernel-based methods, boosting, PAC-Bayesian theory, online
and transductive learning, loss functions, learnable function
classes, notions of complexity for function classes, multitask
learning, and hypothesis selection. These contributions include
historical and context notes, short surveys, and comments on future
research directions. This book will be of interest to researchers,
engineers, and graduate students engaged with all aspects of
statistical learning.
|
Conformal and Probabilistic Prediction with Applications - 5th International Symposium, COPA 2016, Madrid, Spain, April 20-22, 2016, Proceedings (Paperback, 1st ed. 2016)
Alexander Gammerman, Zhiyuan Luo, Jesus Vega, Vladimir Vovk
|
R2,205
Discovery Miles 22 050
|
Ships in 10 - 15 working days
|
This book constitutes the refereed proceedings of the 5th
International Symposium on Conformal and Probabilistic Prediction
with Applications, COPA 2016, held in Madrid, Spain, in April 2016.
The 14 revised full papers presented together with 1 invited paper
were carefully reviewed and selected from 23 submissions and cover
topics on theory of conformal prediction; applications of conformal
prediction; and machine learning.
|
Statistical Learning and Data Sciences - Third International Symposium, SLDS 2015, Egham, UK, April 20-23, 2015, Proceedings (Paperback, 2015 ed.)
Alexander Gammerman, Vladimir Vovk, Harris Papadopoulos
|
R2,841
Discovery Miles 28 410
|
Ships in 10 - 15 working days
|
This book constitutes the refereed proceedings of the Third
International Symposium on Statistical Learning and Data Sciences,
SLDS 2015, held in Egham, Surrey, UK, April 2015. The 36 revised
full papers presented together with 2 invited papers were carefully
reviewed and selected from 59 submissions. The papers are organized
in topical sections on statistical learning and its applications,
conformal prediction and its applications, new frontiers in data
analysis for nuclear fusion, and geometric data analysis.
|
|