![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Computer software packages > Other software packages > Mathematical & statistical software
Data mining is the art and science of intelligent data analysis. By building knowledge from information, data mining adds considerable value to the ever increasing stores of electronic data that abound today. In performing data mining many decisions need to be made regarding the choice of methodology, the choice of data, the choice of tools, and the choice of algorithms. Throughout this book the reader is introduced to the basic concepts and some of the more popular algorithms of data mining. With a focus on the hands-on end-to-end process for data mining, Williams guides the reader through various capabilities of the easy to use, free, and open source Rattle Data Mining Software built on the sophisticated R Statistical Software. The focus on doing data mining rather than just reading about data mining is refreshing. The book covers data understanding, data preparation, data refinement, model building, model evaluation, and practical deployment. The reader will learn to rapidly deliver a data mining project using software easily installed for free from the Internet. Coupling Rattle with R delivers a very sophisticated data mining environment with all the power, and more, of the many commercial offerings.
This book is a text for a one-semester course for upper-level undergraduates and beginning graduate students in engineering, science, and mathematics. Prerequisites are a first course in the theory of ODEs and a survey course in numerical analysis, in addition to specific programming experience, preferably in MATLAB, and knowledge of elementary matrix theory. Professionals will also find that this useful concise reference contains reviews of technical issues and realistic and detailed examples. The programs for the examples are supplied on the accompanying web site and can serve as templates for solving other problems. Each chapter begins with a discussion of the "facts of life" for the problem, mainly by means of examples. Numerical methods for the problem are then developed, but only those methods most widely used. The treatment of each method is brief and technical issues are minimized, but all the issues important in practice and for understaning the codes are discussed. The last part of each chapter is a tutorial that shows how to solve problems by means of small, but realistic, examples.
We have sold 4300 copies worldwide of the first edition (1999). This new edition contains five completely new chapters covering new developments.
The first MATLAB-based numerical methods textbook for bioengineers that uniquely integrates modelling concepts with statistical analysis, while maintaining a focus on enabling the user to report the error or uncertainty in their result. Between traditional numerical method topics of linear modelling concepts, nonlinear root finding, and numerical integration, chapters on hypothesis testing, data regression and probability are interweaved. A unique feature of the book is the inclusion of examples from clinical trials and bioinformatics, which are not found in other numerical methods textbooks for engineers. With a wealth of biomedical engineering examples, case studies on topical biomedical research, and the inclusion of end of chapter problems, this is a perfect core text for a one-semester undergraduate course.
Recommended by Bill Gates A thought-provoking and wide-ranging exploration of machine learning and the race to build computer intelligences as flexible as our own In the world's top research labs and universities, the race is on to invent the ultimate learning algorithm: one capable of discovering any knowledge from data, and doing anything we want, before we even ask. In The Master Algorithm, Pedro Domingos lifts the veil to give us a peek inside the learning machines that power Google, Amazon, and your smartphone. He assembles a blueprint for the future universal learner--the Master Algorithm--and discusses what it will mean for business, science, and society. If data-ism is today's philosophy, this book is its bible.
There has been dramatic growth in the development and application of Bayesian inference in statistics. Berger (2000) documents the increase in Bayesian activity by the number of published research articles, the number of books, andtheextensivenumberofapplicationsofBayesianarticlesinapplied disciplines such as science and engineering. One reason for the dramatic growth in Bayesian modeling is the availab- ity of computational algorithms to compute the range of integrals that are necessary in a Bayesian posterior analysis. Due to the speed of modern c- puters, it is now possible to use the Bayesian paradigm to 't very complex models that cannot be 't by alternative frequentist methods. To 't Bayesian models, one needs a statistical computing environment. This environment should be such that one can: write short scripts to de?ne a Bayesian model use or write functions to summarize a posterior distribution use functions to simulate from the posterior distribution construct graphs to illustrate the posterior inference An environment that meets these requirements is the R system. R provides a wide range of functions for data manipulation, calculation, and graphical d- plays. Moreover, it includes a well-developed, simple programming language that users can extend by adding new functions. Many such extensions of the language in the form of packages are easily downloadable from the Comp- hensive R Archive Network (CRAN)
Computational techniques based on simulation have now become an essential part of the statistician's toolbox. It is thus crucial to provide statisticians with a practical understanding of those methods, and there is no better way to develop intuition and skills for simulation than to use simulation to solve statistical problems. Introducing Monte Carlo Methods with R covers the main tools used in statistical simulation from a programmer's point of view, explaining the R implementation of each simulation technique and providing the output for better understanding and comparison. While this book constitutes a comprehensive treatment of simulation methods, the theoretical justification of those methods has been considerably reduced, compared with Robert and Casella (2004). Similarly, the more exploratory and less stable solutions are not covered here. This book does not require a preliminary exposure to the R programming language or to Monte Carlo methods, nor an advanced mathematical background. While many examples are set within a Bayesian framework, advanced expertise in Bayesian statistics is not required. The book covers basic random generation algorithms, Monte Carlo techniques for integration and optimization, convergence diagnoses, Markov chain Monte Carlo methods, including Metropolis {Hastings and Gibbs algorithms, and adaptive algorithms. All chapters include exercises and all R programs are available as an R package called mcsm. The book appeals to anyone with a practical interest in simulation methods but no previous exposure. It is meant to be useful for students and practitioners in areas such as statistics, signal processing, communications engineering, control theory, econometrics, finance and more. The programming parts are introduced progressively to be accessible to any reader.
This book was first published in 2003. Combinatorica, an extension to the popular computer algebra system Mathematica (R), is the most comprehensive software available for teaching and research applications of discrete mathematics, particularly combinatorics and graph theory. This book is the definitive reference/user's guide to Combinatorica, with examples of all 450 Combinatorica functions in action, along with the associated mathematical and algorithmic theory. The authors cover classical and advanced topics on the most important combinatorial objects: permutations, subsets, partitions, and Young tableaux, as well as all important areas of graph theory: graph construction operations, invariants, embeddings, and algorithmic graph theory. In addition to being a research tool, Combinatorica makes discrete mathematics accessible in new and exciting ways to a wide variety of people, by encouraging computational experimentation and visualization. The book contains no formal proofs, but enough discussion to understand and appreciate all the algorithms and theorems it contains.
How to Use SPSS (R) is designed with the novice computer user in mind and for people who have no previous experience using SPSS. Each chapter is divided into short sections that describe the statistic being used, important underlying assumptions, and how to interpret the results and express them in a research report. The book begins with the basics, such as starting SPSS, defining variables, and entering and saving data. It covers all major statistical techniques typically taught in beginning statistics classes, such as descriptive statistics, graphing data, prediction and association, parametric inferential statistics, nonparametric inferential statistics and statistics for test construction. More than 270 screenshots (including sample output) throughout the book show students exactly what to expect as they follow along using SPSS. The book includes a glossary of statistical terms and practice exercises. A complete set of online resources including video tutorials and output files for students, and PowerPoint slides and test bank questions for instructors, make How to Use SPSS (R) the definitive, field-tested resource for learning SPSS. New to this edition: Now in full color with additional screenshots Fully updated to the reflect SPSS version 26 (and prior versions) Changes in nonparametric tests Model View incorporated Data and real output are now available for all Phrasing Results sections - eliminating hypothetical output or hypothetical data
"Modeling with Data" fully explains how to execute computationally intensive analyses on very large data sets, showing readers how to determine the best methods for solving a variety of different problems, how to create and debug statistical models, and how to run an analysis and evaluate the results. Ben Klemens introduces a set of open and unlimited tools, and uses them to demonstrate data management, analysis, and simulation techniques essential for dealing with large data sets and computationally intensive procedures. He then demonstrates how to easily apply these tools to the many threads of statistical technique, including classical, Bayesian, maximum likelihood, and Monte Carlo methods. Klemens's accessible survey describes these models in a unified and nontraditional manner, providing alternative ways of looking at statistical concepts that often befuddle students. The book includes nearly one hundred sample programs of all kinds. Links to these programs will be available on this page at a later date. "Modeling with Data" will interest anyone looking for a comprehensive guide to these powerful statistical tools, including researchers and graduate students in the social sciences, biology, engineering, economics, and applied mathematics.
R is a language and environment for data analysis and graphics. It may be considered an implementation of S, an award-winning language initially - veloped at Bell Laboratories since the late 1970s. The R project was initiated by Robert Gentleman and Ross Ihaka at the University of Auckland, New Zealand, in the early 1990s, and has been developed by an international team since mid-1997. Historically, econometricians have favored other computing environments, some of which have fallen by the wayside, and also a variety of packages with canned routines. We believe that R has great potential in econometrics, both for research and for teaching. There are at least three reasons for this: (1) R is mostly platform independent and runs on Microsoft Windows, the Mac family of operating systems, and various ?avors of Unix/Linux, and also on some more exotic platforms. (2) R is free software that can be downloaded and installed at no cost from a family of mirror sites around the globe, the Comprehensive R Archive Network (CRAN); hence students can easily install it on their own machines. (3) R is open-source software, so that the full source code is available and can be inspected to understand what it really does, learn from it, and modify and extend it. We also like to think that platform independence and the open-source philosophy make R an ideal environment for reproducible econometric research.
This volume contains selected papers presented at two joint German-Japanese symposia on data analysis and related elds. The articles substantially extend and further develop material presented at the two symposia organized on the basis of longstanding and close relationships which have been cultivated in the last couple of decades between the two classi cation societies: the German Class- cation Society (Gesellschaft fu ]r Klassi kation e. V.) and the Japanese Classi cation Society. These symposia have been very helpful in exchanging ideas, views, and knowledge between the two societies and have served as a spring board for more extensive and closer co-operation between the societies as well as among their individual members. The scienti c program of the rst Joint Japanese-German Symposium (Tokyo 2005)included23presentations;forthesecondJointGerman-JapaneseSymposium (Berlin 2006) 27 presentations were scheduled. This volume presents 21 peer refereed papers, which are grouped into three parts: 1. Part 1 Clustering and Visualization (eight papers) 2. Part 2 Methods in Fields (nine papers) 3. Part 3 Applications in Clustering and Visualization (four papers) The concept of having a joint symposium of the two classi cation societies came from the talks with Hans-Hermann and Wolfgang when Akinori attended the 28th Annual Conference of the German Classi cation Society held in Dortmund in March 2004."
"Mathematica Cookbook" helps you master the application's core principles by walking you through real-world problems. Ideal for browsing, this book includes recipes for working with numerics, data structures, algebraic equations, calculus, and statistics. You'll also venture into exotic territory with recipes for data visualization using 2D and 3D graphic tools, image processing, and music. Although Mathematica 7 is a highly advanced computational platform, the recipes in this book make it accessible to everyone -- whether you're working on high school algebra, simple graphs, PhD-level computation, financial analysis, or advanced engineering models.Learn how to use Mathematica at a higher level with functional programming and pattern matchingDelve into the rich library of functions for string and structured text manipulationLearn how to apply the tools to physics and engineering problemsDraw on Mathematica's access to physics, chemistry, and biology dataGet techniques for solving equations in computational financeLearn how to use Mathematica for sophisticated image processingProcess music and audio as musical notes, analog waveforms, or digital sound samples
An elementary first course for students in mathematics and engineering Practical in approach: examples of code are provided for students to debug, and tasks - with full solutions - are provided at the end of each chapter Includes a glossary of useful terms, with each term supported by an example of the syntaxes commonly encountered
New statistical tools are changing the ways in which scientists analyze and interpret data and models. Many of these are emerging as a result of the wide availability of inexpensive, high speed computational power. In particular, hierarchical Bayes and Markov Chain Monte Carlo methods for anyalsis provide consistent framework for inference and prediction where information is heterogeneous and uncertain, processes are complex, and responses depend on scale. Nowhere are these methods more promising than in the environmental sciences. Models have developed rapidly, and there is now a requirement for a clear exposition of the methodology through to application for a range of environmental challenges.
The second edition of this popular guide demonstrates the process of entering and analyzing data using the latest version of SPSS (12.0), and is also appropriate for those using earlier versions of SPSS. The book is easy to follow because all procedures are outlined in a step-by-step format designed for the novice user. Students are introduced to the rationale of statistical tests and detailed explanations of results are given through clearly annotated examples of SPSS output. Topics covered range from descriptive statistics through multiple regression analysis. In addition, this guide includes topics not typically covered in other books such as probability theory, interaction effects in analysis of variance, factor analysis, and scale reliability. Chapter exercises reinforce the text examples and may be performed for further practice, for homework assignments, or in computer laboratory sessions. This book can be used in two ways: as a stand-alone manual for students wishing to learn data analysis techniques using SPSS for Windows, or in research and statistics courses to be used with a basic statistics text. The book provides hands-on experience with actual data sets, helps students choose appropriate statistical tests, illustrates the meaning of results, and provides exercises to be completed for further practice or as homework assignments. Susan B. Gerber, Ph.D. is Research Assistant Professor of Education at State University of New York at Buffalo. She is director of the Educational Technology program and holds degrees in Statistics and Educational Psychology. Kristin Voelkl Finn, Ph.D. is Assistant Professor of Education at Canisius College. She teaches graduate courses in research methodology and conducts research on adolescent problem behavior.
A self-contained introduction to abstract interpretation-based static analysis, an essential resource for students, developers, and users. Static program analysis, or static analysis, aims to discover semantic properties of programs without running them. It plays an important role in all phases of development, including verification of specifications and programs, the synthesis of optimized code, and the refactoring and maintenance of software applications. This book offers a self-contained introduction to static analysis, covering the basics of both theoretical foundations and practical considerations in the use of static analysis tools. By offering a quick and comprehensive introduction for nonspecialists, the book fills a notable gap in the literature, which until now has consisted largely of scientific articles on advanced topics. The text covers the mathematical foundations of static analysis, including semantics, semantic abstraction, and computation of program invariants; more advanced notions and techniques, including techniques for enhancing the cost-accuracy balance of analysis and abstractions for advanced programming features and answering a wide range of semantic questions; and techniques for implementing and using static analysis tools. It begins with background information and an intuitive and informal introduction to the main static analysis principles and techniques. It then formalizes the scientific foundations of program analysis techniques, considers practical aspects of implementation, and presents more advanced applications. The book can be used as a textbook in advanced undergraduate and graduate courses in static analysis and program verification, and as a reference for users, developers, and experts.
Most biologists use nonlinear regression more than any other statistical technique, but there are very few places to learn about curve-fitting. This book, by the author of the very successful Intuitive Biostatistics, addresses this relatively focused need of an extraordinarily broad range of scientists.
This volume brings together the main results in the field of Bayesian Optimization (BO), focusing on the last ten years and showing how, on the basic framework, new methods have been specialized to solve emerging problems from machine learning, artificial intelligence, and system optimization. It also analyzes the software resources available for BO and a few selected application areas. Some areas for which new results are shown include constrained optimization, safe optimization, and applied mathematics, specifically BO's use in solving difficult nonlinear mixed integer problems. The book will help bring readers to a full understanding of the basic Bayesian Optimization framework and gain an appreciation of its potential for emerging application areas. It will be of particular interest to the data science, computer science, optimization, and engineering communities.
Thoroughly updated with improved pedagogy, the fifth edition of this classic textbook continues to provide students with a clear and comprehensive introduction the fundamentals of the finite element method. New features include enhanced coverage of introductory topics in the context of simple 1D problems, providing students with a solid base from which to advance to 2D and 3D problems; expanded coverage of more advanced concepts, to reinforce students' understanding; over 30 additional solved problems; and downloadable MATLAB, Python, C, Javascript, Fortran and Excel VBA code packages, providing students with hands-on experience, and preparing them for commercial software. Accompanied by online solutions for instructors, this is the definitive text for senior undergraduate and graduate students studying a first course in the finite element method and finite element analysis, and for professional engineers keen to shore up their understanding of finite element fundamentals.
Reproducible Finance with R: Code Flows and Shiny Apps for Portfolio Analysis is a unique introduction to data science for investment management that explores the three major R/finance coding paradigms, emphasizes data visualization, and explains how to build a cohesive suite of functioning Shiny applications. The full source code, asset price data and live Shiny applications are available at reproduciblefinance.com. The ideal reader works in finance or wants to work in finance and has a desire to learn R code and Shiny through simple, yet practical real-world examples. The book begins with the first step in data science: importing and wrangling data, which in the investment context means importing asset prices, converting to returns, and constructing a portfolio. The next section covers risk and tackles descriptive statistics such as standard deviation, skewness, kurtosis, and their rolling histories. The third section focuses on portfolio theory, analyzing the Sharpe Ratio, CAPM, and Fama French models. The book concludes with applications for finding individual asset contribution to risk and for running Monte Carlo simulations. For each of these tasks, the three major coding paradigms are explored and the work is wrapped into interactive Shiny dashboards.
This book brings together some of the finest minds in the statistical and neural network research communities. It provides a broad overview of important current developments and highlights the likely future trends in the area of neural networks.
Gain the R programming language fundamentals for doing the applied statistics useful for data exploration and analysis in data science and data mining. This book covers topics ranging from R syntax basics, descriptive statistics, and data visualizations to inferential statistics and regressions. After learning R's syntax, you will work through data visualizations such as histograms and boxplot charting, descriptive statistics, and inferential statistics such as t-test, chi-square test, ANOVA, non-parametric test, and linear regressions. Learn R for Applied Statistics is a timely skills-migration book that equips you with the R programming fundamentals and introduces you to applied statistics for data explorations. What You Will Learn Discover R, statistics, data science, data mining, and big data Master the fundamentals of R programming, including variables and arithmetic, vectors, lists, data frames, conditional statements, loops, and functions Work with descriptive statistics Create data visualizations, including bar charts, line charts, scatter plots, boxplots, histograms, and scatterplots Use inferential statistics including t-tests, chi-square tests, ANOVA, non-parametric tests, linear regressions, and multiple linear regressions Who This Book Is For Those who are interested in data science, in particular data exploration using applied statistics, and the use of R programming for data visualizations. |
![]() ![]() You may like...
Recent Trends in Mechanical Engineering…
C. S. Ramesh, Praduymna Ghosh, …
Hardcover
R6,326
Discovery Miles 63 260
Regulation and the Evolution of the…
Anastassios Gentzoglanis, Anders Henten
Hardcover
R4,028
Discovery Miles 40 280
The Unicode Cookbook for Linguists
Steven Moran, Michael Cysouw
Hardcover
R1,059
Discovery Miles 10 590
Programming Environments for Massively…
K.M. Decker, R. M Rehmann
Hardcover
R2,604
Discovery Miles 26 040
|