|
Showing 1 - 4 of
4 matches in All Departments
A Computational Approach to Statistical Learning gives a novel
introduction to predictive modeling by focusing on the algorithmic
and numeric motivations behind popular statistical methods. The
text contains annotated code to over 80 original reference
functions. These functions provide minimal working implementations
of common statistical learning algorithms. Every chapter concludes
with a fully worked out application that illustrates predictive
modeling tasks using a real-world dataset. The text begins with a
detailed analysis of linear models and ordinary least squares.
Subsequent chapters explore extensions such as ridge regression,
generalized linear models, and additive models. The second half
focuses on the use of general-purpose algorithms for convex
optimization and their application to tasks in statistical
learning. Models covered include the elastic net, dense neural
networks, convolutional neural networks (CNNs), and spectral
clustering. A unifying theme throughout the text is the use of
optimization theory in the description of predictive models, with a
particular focus on the singular value decomposition (SVD). Through
this theme, the computational approach motivates and clarifies the
relationships between various predictive models. Taylor Arnold is
an assistant professor of statistics at the University of Richmond.
His work at the intersection of computer vision, natural language
processing, and digital humanities has been supported by multiple
grants from the National Endowment for the Humanities (NEH) and the
American Council of Learned Societies (ACLS). His first book,
Humanities Data in R, was published in 2015. Michael Kane is an
assistant professor of biostatistics at Yale University. He is the
recipient of grants from the National Institutes of Health (NIH),
DARPA, and the Bill and Melinda Gates Foundation. His R package
bigmemory won the Chamber's prize for statistical software in 2010.
Bryan Lewis is an applied mathematician and author of many popular
R packages, including irlba, doRedis, and threejs.
A Computational Approach to Statistical Learning gives a novel
introduction to predictive modeling by focusing on the algorithmic
and numeric motivations behind popular statistical methods. The
text contains annotated code to over 80 original reference
functions. These functions provide minimal working implementations
of common statistical learning algorithms. Every chapter concludes
with a fully worked out application that illustrates predictive
modeling tasks using a real-world dataset. The text begins with a
detailed analysis of linear models and ordinary least squares.
Subsequent chapters explore extensions such as ridge regression,
generalized linear models, and additive models. The second half
focuses on the use of general-purpose algorithms for convex
optimization and their application to tasks in statistical
learning. Models covered include the elastic net, dense neural
networks, convolutional neural networks (CNNs), and spectral
clustering. A unifying theme throughout the text is the use of
optimization theory in the description of predictive models, with a
particular focus on the singular value decomposition (SVD). Through
this theme, the computational approach motivates and clarifies the
relationships between various predictive models. Taylor Arnold is
an assistant professor of statistics at the University of Richmond.
His work at the intersection of computer vision, natural language
processing, and digital humanities has been supported by multiple
grants from the National Endowment for the Humanities (NEH) and the
American Council of Learned Societies (ACLS). His first book,
Humanities Data in R, was published in 2015. Michael Kane is an
assistant professor of biostatistics at Yale University. He is the
recipient of grants from the National Institutes of Health (NIH),
DARPA, and the Bill and Melinda Gates Foundation. His R package
bigmemory won the Chamber's prize for statistical software in 2010.
Bryan Lewis is an applied mathematician and author of many popular
R packages, including irlba, doRedis, and threejs.
This pioneering book teaches readers to use R within four core
analytical areas applicable to the Humanities: networks, text,
geospatial data, and images. This book is also designed to be a
bridge: between quantitative and qualitative methods, individual
and collaborative work, and the humanities and social sciences.
Humanities Data with R does not presuppose background programming
experience. Early chapters take readers from R set-up to
exploratory data analysis (continuous and categorical data,
multivariate analysis, and advanced graphics with emphasis on
aesthetics and facility). Following this, networks, geospatial
data, image data, natural language processing and text analysis
each have a dedicated chapter. Each chapter is grounded in examples
to move readers beyond the intimidation of adding new tools to
their research. Everything is hands-on: networks are explained
using U.S. Supreme Court opinions, and low-level NLP methods are
applied to short stories by Sir Arthur Conan Doyle. After working
through these examples with the provided data, code and book
website, readers are prepared to apply new methods to their own
work. The open source R programming language, with its myriad
packages and popularity within the sciences and social sciences, is
particularly well-suited to working with humanities data. R
packages are also highlighted in an appendix. This book uses an
expanded conception of the forms data may take and the information
it represents. The methodology will have wide application in
classrooms and self-study for the humanities, but also for use in
linguistics, anthropology, and political science. Outside the
classroom, this intersection of humanities and computing is
particularly relevant for research and new modes of dissemination
across archives, museums and libraries.
|
You may like...
Loot
Nadine Gordimer
Paperback
(2)
R383
R310
Discovery Miles 3 100
|