![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Computer software packages > Other software packages > Mathematical & statistical software
Praise for the First Edition "The main strength of this book is that it provides a unified framework of graphical tools for data analysis, especially for univariate and low-dimensional multivariate data. In addition, it is clearly written in plain language and the inclusion of R code is particularly useful to assist readers' understanding of the graphical techniques discussed in the book. ... It not only summarises graphical techniques, but it also serves as a practical reference for researchers and graduate students with an interest in data display." -Han Lin Shang, Journal of Applied Statistics Graphics for Statistics and Data Analysis with R, Second Edition, presents the basic principles of graphical design and applies these principles to engaging examples using the graphics and lattice packages in R. It offers a wide array of modern graphical displays for data visualization and representation. Added in the second edition are coverage of the ggplot2 graphics package, material on human visualization and color rendering in R, on screen, and in print. Features Emphasizes the fundamentals of statistical graphics and best practice guidelines for producing and choosing among graphical displays in R Presents technical details on topics such as: the estimation of quantiles, nonparametric and parametric density estimation; diagnostic plots for the simple linear regression model; polynomial regression, splines, and locally weighted polynomial regression for producing a smooth curve; Trellis graphics for multivariate data Provides downloadable R code and data for figures at www.graphicsforstatistics.com Kevin J. Keen is a Professor of Mathematics and Statistics at the University of Northern British Columbia (Prince George, Canada) and an Accredited Professional StatisticianTM by the Statistical Society of Canada and the American Statistical Association.
Mathematica by Example, Sixth Edition is an essential resource for the Mathematica user, providing step-by-step instructions on achieving results from this powerful software tool. The book fully accounts for the changes to functionality and visualization capabilities and accomodates the full array of new extensions in the types of data and problems that Mathematica can immediately handle, including cloud services and systems, geographic and geometric computation, dynamic visualization, interactive applications and other improvements. It is an ideal text for scientific students, researchers, and aspiring programmers seeking further understanding of Mathematica. Written by seasoned practitioners with a view to practical implementation and problem-solving, the book's pedagogy is delivered clearly and without jargon using representative biological, physical and engineering problems. Code is provided on an ancillary website to support the use of Mathematica across diverse applications and subject areas.
Simple English format Foundation sections on inference and evidence, and data integration in project management Exploration of R usage through a narrative examining a generic integrated conservation and development project A final section on R for reproducible workflow Accompanied by an R package
Big Data: A Tutorial-Based Approach explores the tools and techniques used to bring about the marriage of structured and unstructured data. It focuses on Hadoop Distributed Storage and MapReduce Processing by implementing (i) Tools and Techniques of Hadoop Eco System, (ii) Hadoop Distributed File System Infrastructure, and (iii) efficient MapReduce processing. The book includes Use Cases and Tutorials to provide an integrated approach that answers the 'What', 'How', and 'Why' of Big Data. Features Identifies the primary drivers of Big Data Walks readers through the theory, methods and technology of Big Data Explains how to handle the 4 V's of Big Data in order to extract value for better business decision making Shows how and why data connectors are critical and necessary for Agile text analytics Includes in-depth tutorials to perform necessary set-ups, installation, configuration and execution of important tasks Explains the command line as well as GUI interface to a powerful data exchange tool between Hadoop and legacy r-dbms databases
Praise for the first edition: [This book] reflects the extensive experience and significant contributions of the author to non-linear and non-Gaussian modeling. ... [It] is a valuable book, especially with its broad and accessible introduction of models in the state-space framework. -Statistics in Medicine What distinguishes this book from comparable introductory texts is the use of state-space modeling. Along with this come a number of valuable tools for recursive filtering and smoothing, including the Kalman filter, as well as non-Gaussian and sequential Monte Carlo filters. -MAA Reviews Introduction to Time Series Modeling with Applications in R, Second Edition covers numerous stationary and nonstationary time series models and tools for estimating and utilizing them. The goal of this book is to enable readers to build their own models to understand, predict and master time series. The second edition makes it possible for readers to reproduce examples in this book by using the freely available R package TSSS to perform computations for their own real-world time series problems. This book employs the state-space model as a generic tool for time series modeling and presents the Kalman filter, the non-Gaussian filter and the particle filter as convenient tools for recursive estimation for state-space models. Further, it also takes a unified approach based on the entropy maximization principle and employs various methods of parameter estimation and model selection, including the least squares method, the maximum likelihood method, recursive estimation for state-space models and model selection by AIC. Along with the standard stationary time series models, such as the AR and ARMA models, the book also introduces nonstationary time series models such as the locally stationary AR model, the trend model, the seasonal adjustment model, the time-varying coefficient AR model and nonlinear non-Gaussian state-space models. About the Author: Genshiro Kitagawa is a project professor at the University of Tokyo, the former Director-General of the Institute of Statistical Mathematics, and the former President of the Research Organization of Information and Systems.
Modeling spatial and spatio-temporal continuous processes is an important and challenging problem in spatial statistics. Advanced Spatial Modeling with Stochastic Partial Differential Equations Using R and INLA describes in detail the stochastic partial differential equations (SPDE) approach for modeling continuous spatial processes with a Matern covariance, which has been implemented using the integrated nested Laplace approximation (INLA) in the R-INLA package. Key concepts about modeling spatial processes and the SPDE approach are explained with examples using simulated data and real applications. This book has been authored by leading experts in spatial statistics, including the main developers of the INLA and SPDE methodologies and the R-INLA package. It also includes a wide range of applications: * Spatial and spatio-temporal models for continuous outcomes * Analysis of spatial and spatio-temporal point patterns * Coregionalization spatial and spatio-temporal models * Measurement error spatial models * Modeling preferential sampling * Spatial and spatio-temporal models with physical barriers * Survival analysis with spatial effects * Dynamic space-time regression * Spatial and spatio-temporal models for extremes * Hurdle models with spatial effects * Penalized Complexity priors for spatial models All the examples in the book are fully reproducible. Further information about this book, as well as the R code and datasets used, is available from the book website at http://www.r-inla.org/spde-book. The tools described in this book will be useful to researchers in many fields such as biostatistics, spatial statistics, environmental sciences, epidemiology, ecology and others. Graduate and Ph.D. students will also find this book and associated files a valuable resource to learn INLA and the SPDE approach for spatial modeling.
This unique book addresses the statistical modelling and analysis of microbiome data using cutting-edge R software. It includes real-world data from the authors' research and from the public domain, and discusses the implementation of R for data analysis step by step. The data and R computer programs are publicly available, allowing readers to replicate the model development and data analysis presented in each chapter, so that these new methods can be readily applied in their own research. The book also discusses recent developments in statistical modelling and data analysis in microbiome research, as well as the latest advances in next-generation sequencing and big data in methodological development and applications. This timely book will greatly benefit all readers involved in microbiome, ecology and microarray data analyses, as well as other fields of research.
R Visualizations: Derive Meaning from Data focuses on one of the two major topics of data analytics: data visualization, a.k.a., computer graphics. In the book, major R systems for visualization are discussed, organized by topic and not by system. Anyone doing data analysis will be shown how to use R to generate any of the basic visualizations with the R visualization systems. Further, this book introduces the author's lessR system, which always can accomplish a visualization with less coding than the use of other systems, sometimes dramatically so, and also provides accompanying statistical analyses. Key Features Presents thorough coverage of the leading R visualization system, ggplot2. Gives specific guidance on using base R graphics to attain visualizations of the same quality as those provided by ggplot2. Shows how to create a wide range of data visualizations: distributions of categorical and continuous variables, many types of scatterplots including with a third variable, time series, and maps. Inclusion of the various approaches to R graphics organized by topic instead of by system. Presents the recent work on interactive visualization in R. David W. Gerbing received his PhD from Michigan State University in 1979 in quantitative analysis, and currently is a professor of quantitative analysis in the School of Business at Portland State University. He has published extensively in the social and behavioral sciences with a focus on quantitative methods. His lessR package has been in development since 2009.
Technological improvements continue to push back the frontier of processor speed in modern computers. Unfortunately, the computational intensity demanded by modern research problems grows even faster. Parallel computing has emerged as the most successful bridge to this computational gap, and many popular solutions have emerged based on its concepts, such as grid computing and massively parallel supercomputers. The Handbook of Parallel Computing and Statistics systematically applies the principles of parallel computing for solving increasingly complex problems in statistics research. This unique reference weaves together the principles and theoretical models of parallel computing with the design, analysis, and application of algorithms for solving statistical problems. After a brief introduction to parallel computing, the book explores the architecture, programming, and computational aspects of parallel processing. Focus then turns to optimization methods followed by statistical applications. These applications include algorithms for predictive modeling, adaptive design, real-time estimation of higher-order moments and cumulants, data mining, econometrics, and Bayesian computation. Expert contributors summarize recent results and explore new directions in these areas. Its intricate combination of theory and practical applications makes the Handbook of Parallel Computing and Statistics an ideal companion for helping solve the abundance of computation-intensive statistical problems arising in a variety of fields.
This volume features original contributions and invited review articles on mathematical statistics, statistical simulation and experimental design. The selected peer-reviewed contributions originate from the 8th International Workshop on Simulation held in Vienna in 2015. The book is intended for mathematical statisticians, Ph.D. students and statisticians working in medicine, engineering, pharmacy, psychology, agriculture and other related fields. The International Workshops on Simulation are devoted to statistical techniques in stochastic simulation, data collection, design of scientific experiments and studies representing broad areas of interest. The first 6 workshops took place in St. Petersburg, Russia, in 1994 - 2009 and the 7th workshop was held in Rimini, Italy, in 2013.
A hands-on guide to using Excel in the business context First published in 2012, Using Excel for Business and Financial Modelling contains step-by-step instructions of how to solve common business problems using financial models, including downloadable Excel templates, a list of shortcuts and tons of practical tips and techniques you can apply straight away. Whilst there are many hundreds of tools, features and functions in Excel, this book focuses on the topics most relevant to finance professionals. It covers these features in detail from a practical perspective, but also puts them in context by applying them to practical examples in the real world. Learn to create financial models to help make business decisions whilst applying modelling best practice methodology, tools and techniques. - Provides the perfect mix of practice and theory - Helps you become a DIY Excel modelling specialist - Includes updates for Excel 2019/365 and Excel for Mac - May be used as an accompaniment to the author's online and face-to-face training courses Many people are often overwhelmed by the hundreds of tools in Excel, and this book gives clarity to the ones you need to know in order to perform your job more efficiently. This book also demystifies the technical, design, logic and financial skills you need for business and financial modelling.
This workbook and solutions manual is intended for advanced undergraduate or beginning graduate students as a supplement to a traditional course in numerical mathematics and as preparation for independent research involving numerical mathematics. The solutions manual provides complete MATLAB code and numerical results for each of the exercises in the workbook and will be especially useful for those students without previous MATLAB programming experience. It is also valuable for classroom instructors to help pinpoint the author's intent in each exercise and to provide a model for graders. Upon completion of this material, students will have a working knowledge of MATLAB programming, they will have themselves programmed algorithms encountered in classwork and textbooks, and they will know how to check and verify their own programs against hand calculations and by reference to theoretical results, special polynomial solutions and other specialized solutions. No previous programming experience with MATLAB is necessary.
Interactive Graphics for Data Analysis: Principles and Examples discusses exploratory data analysis (EDA) and how interactive graphical methods can help gain insights as well as generate new questions and hypotheses from datasets. Fundamentals of Interactive Statistical GraphicsThe first part of the book summarizes principles and methodology, demonstrating how the different graphical representations of variables of a dataset are effectively used in an interactive setting. The authors introduce the most important plots and their interactive controls. They also examine various types of data, relations between variables, and plot ensembles. Case Studies Illustrate the PrinciplesThe second section focuses on nine case studies. Each case study describes the background, lists the main goals of the analysis and the variables in the dataset, shows what further numerical procedures can add to the graphical analysis, and summarizes important findings. Wherever applicable, the authors also provide the numerical analysis for datasets found in Cox and Snell's landmark book. Understand How to Analyze Data through Graphical Means This full-color text shows that interactive graphical methods complement the traditional statistical toolbox to achieve more complete, easier to understand, and easier to interpret analyses.
The richly illustrated Interactive Web-Based Data Visualization with R, plotly, and shiny focuses on the process of programming interactive web graphics for multidimensional data analysis. It is written for the data analyst who wants to leverage the capabilities of interactive web graphics without having to learn web programming. Through many R code examples, you will learn how to tap the extensive functionality of these tools to enhance the presentation and exploration of data. By mastering these concepts and tools, you will impress your colleagues with your ability to quickly generate more informative, engaging, and reproducible interactive graphics using free and open source software that you can share over email, export to pdf, and more. Key Features: Convert static ggplot2 graphics to an interactive web-based form Link, animate, and arrange multiple plots in standalone HTML from R Embed, modify, and respond to plotly graphics in a shiny app Learn best practices for visualizing continuous, discrete, and multivariate data Learn numerous ways to visualize geo-spatial data This book makes heavy use of plotly for graphical rendering, but you will also learn about other R packages that support different phases of a data science workflow, such as tidyr, dplyr, and tidyverse. Along the way, you will gain insight into best practices for visualization of high-dimensional data, statistical graphics, and graphical perception. The printed book is complemented by an interactive website where readers can view movies demonstrating the examples and interact with graphics.
This book systematically addresses the design and analysis of efficient techniques for independent random sampling. Both general-purpose approaches, which can be used to generate samples from arbitrary probability distributions, and tailored techniques, designed to efficiently address common real-world practical problems, are introduced and discussed in detail. In turn, the monograph presents fundamental results and methodologies in the field, elaborating and developing them into the latest techniques. The theory and methods are illustrated with a varied collection of examples, which are discussed in detail in the text and supplemented with ready-to-run computer code. The main problem addressed in the book is how to generate independent random samples from an arbitrary probability distribution with the weakest possible constraints or assumptions in a form suitable for practical implementation. The authors review the fundamental results and methods in the field, address the latest methods, and emphasize the links and interplay between ostensibly diverse techniques.
Financial, Macro and Micro Econometrics Using R, Volume 42, provides state-of-the-art information on important topics in econometrics, including multivariate GARCH, stochastic frontiers, fractional responses, specification testing and model selection, exogeneity testing, causal analysis and forecasting, GMM models, asset bubbles and crises, corporate investments, classification, forecasting, nonstandard problems, cointegration, financial market jumps and co-jumps, among other topics.
This book focuses on computer intensive statistical methods, such as validation, model selection, and bootstrap, that help overcome obstacles that could not be previously solved by methods such as regression and time series modelling in the areas of economics, meteorology, and transportation.
"If mathematical modeling is the process of turning real phenomena into mathematical abstractions, then numerical computation is largely about the transformation from abstract mathematics to concrete reality. Many science and engineering disciplines have long benefited from the tremendous value of the correspondence between quantitative information and mathematical manipulation." -from the Preface Fundamentals of Numerical Computation is an advanced undergraduate-level introduction to the mathematics and use of algorithms for the fundamental problems of numerical computation: linear algebra, finding roots, approximating data and functions, and solving differential equations. The book is organized with simpler methods in the first half and more advanced methods in the second half, allowing use for either a single course or a sequence of two courses. The authors take readers from basic to advanced methods, illustrating them with over 200 self-contained MATLAB functions and examples designed for those with no prior MATLAB experience. Although the text provides many examples, exercises, and illustrations, the aim of the authors is not to provide a cookbook per se, but rather an exploration of the principles of cooking. Professors Driscoll and Braun have developed an online resource that includes well-tested materials related to every chapter. Among these materials are lecture-related slides and videos, ideas for student projects, laboratory exercises, computational examples and scripts, and all the functions presented in the book.
Regression Modeling: Methods, Theory, and Computation with SAS provides an introduction to a diverse assortment of regression techniques using SAS to solve a wide variety of regression problems. The author fully documents the SAS programs and thoroughly explains the output produced by the programs. The text presents the popular ordinary least squares (OLS) approach before introducing many alternative regression methods. It covers nonparametric regression, logistic regression (including Poisson regression), Bayesian regression, robust regression, fuzzy regression, random coefficients regression, L1 and q-quantile regression, regression in a spatial domain, ridge regression, semiparametric regression, nonlinear least squares, and time-series regression issues. For most of the regression methods, the author includes SAS procedure code, enabling readers to promptly perform their own regression runs. A Comprehensive, Accessible Source on Regression Methodology and ModelingRequiring only basic knowledge of statistics and calculus, this book discusses how to use regression analysis for decision making and problem solving. It shows readers the power and diversity of regression techniques without overwhelming them with calculations.
Thoroughly revised and updated, The Art of Modeling in Science and Engineering with Mathematica (R), Second Edition explores the mathematical tools and procedures used in modeling based on the laws of conservation of mass, energy, momentum, and electrical charge. The authors have culled and consolidated the best from the first edition and expanded the range of applied examples to reach a wider audience. The text proceeds, in measured steps, from simple models of real-world problems at the algebraic and ordinary differential equations (ODE) levels to more sophisticated models requiring partial differential equations. The traditional solution methods are supplemented with Mathematica , which is used throughout the text to arrive at solutions for many of the problems presented. The text is enlivened with a host of illustrations and practice problems drawn from classical and contemporary sources. They range from Thomson's famous experiment to determine e/m and Euler's model for the buckling of a strut to an analysis of the propagation of emissions and the performance of wind turbines. The mathematical tools required are first explained in separate chapters and then carried along throughout the text to solve and analyze the models. Commentaries at the end of each illustration draw attention to the pitfalls to be avoided and, perhaps most important, alert the reader to unexpected results that defy conventional wisdom. These features and more make the book the perfect tool for resolving three common difficulties: the proper choice of model, the absence of precise solutions, and the need to make suitable simplifying assumptions and approximations. The book covers a wide range of physical processes and phenomena drawn from various disciplines and clearly illuminates the link between the physical system being modeled and the mathematical expression that results.
The main focus of this book is on presenting advances in fuzzy statistics, and on proposing a methodology for testing hypotheses in the fuzzy environment based on the estimation of fuzzy confidence intervals, a context in which not only the data but also the hypotheses are considered to be fuzzy. The proposed method for estimating these intervals is based on the likelihood method and employs the bootstrap technique. A new metric generalizing the signed distance measure is also developed. In turn, the book presents two conceptually diverse applications in which defended intervals play a role: one is a novel methodology for evaluating linguistic questionnaires developed at the global and individual levels; the other is an extension of the multi-ways analysis of variance to the space of fuzzy sets. To illustrate these approaches, the book presents several empirical and simulation-based studies with synthetic and real data sets. In closing, it presents a coherent R package called "FuzzySTs" which covers all the previously mentioned concepts with full documentation and selected use cases. Given its scope, the book will be of interest to all researchers whose work involves advanced fuzzy statistical methods.
Without question, statistics is one of the most challenging courses for students in the social and behavioral sciences. Enrolling in their first statistics course, students are often apprehensive or extremely anxious toward the subject matter. And while IBM SPSS is one of the more easy-to-use statistical software programs available, for anxious students who realize they not only have to learn statistics but also new software, the task can seem insurmountable. Keenly aware of students' anxiety with statistics (and the fact that this anxiety can affect performance), Ronald D. Yockey has written SPSS Demystified: A Simple Guide and Reference, now in its fourth edition. Through a comprehensive, step-by-step approach, this text is consistently and specifically designed to both alleviate anxiety toward the subject matter and build a successful experience analyzing data in SPSS. Topics covered in the text are appropriate for most introductory and intermediate statistics and research methods courses. Key features of the text: Step-by-step instruction and screenshots Designed to be hands-on with the user performing the analyses alongside on their computer as they read through each chapter Call-out boxes provided, highlighting important information as appropriate SPSS output explained, with written results provided using the popular, widely recognized APA format End-of-chapter exercises included, allowing for additional practice SPSS datasets available on the publisher's website New to the Fourth Edition: Fully updated to SPSS 28 Updated screenshots in full color to reflect changes in SPSS software system (version 28) Exercises updated with up-to-date examples Exact p-values provided (consist with APA recommendations)
The Second Bayesian Young Statisticians Meeting (BAYSM 2014) and the research presented here facilitate connections among researchers using Bayesian Statistics by providing a forum for the development and exchange of ideas. WU Vienna University of Business and Economics hosted BAYSM 2014 from September 18th to the 19th. The guidance of renowned plenary lecturers and senior discussants is a critical part of the meeting and this volume, which follows publication of contributions from BAYSM 2013. The meeting's scientific program reflected the variety of fields in which Bayesian methods are currently employed or could be introduced in the future. Three brilliant keynote lectures by Chris Holmes (University of Oxford), Christian Robert (Universite Paris-Dauphine), and Mike West (Duke University), were complemented by 24 plenary talks covering the major topics Dynamic Models, Applications, Bayesian Nonparametrics, Biostatistics, Bayesian Methods in Economics, and Models and Methods, as well as a lively poster session with 30 contributions. Selected contributions have been drawn from the conference for this book. All contributions in this volume are peer-reviewed and share original research in Bayesian computation, application, and theory.
Compositional Data Analysis in Practice is a user-oriented practical guide to the analysis of data with the property of a constant sum, for example percentages adding up to 100%. Compositional data can give misleading results if regular statistical methods are applied, and are best analysed by first transforming them to logarithms of ratios. This book explains how this transformation affects the analysis, results and interpretation of this very special type of data. All aspects of compositional data analysis are considered: visualization, modelling, dimension-reduction, clustering and variable selection, with many examples in the fields of food science, archaeology, sociology and biochemistry, and a final chapter containing a complete case study using fatty acid compositions in ecology. The applicability of these methods extends to other fields such as linguistics, geochemistry, marketing, economics and finance. R Software The following repository contains data files and R scripts from the book https://github.com/michaelgreenacre/CODAinPractice. The R package easyCODA, which accompanies this book, is available on CRAN -- note that you should have version 0.25 or higher. The latest version of the package will always be available on R-Forge and can be installed from R with this instruction: install.packages("easyCODA", repos="http://R-Forge.R-project.org"). |
![]() ![]() You may like...
Blast Mitigation - Experimental and…
Arun Shukla, Yapa D.S. Rajapakse, …
Hardcover
R5,121
Discovery Miles 51 210
Geospatial Algebraic Computations…
Joseph Awange, Bela Palancz
Hardcover
Semiconductor Modeling: - For Simulating…
Roy Leventhal
Hardcover
Biomedical Engineering - Health Care…
Sang C Suh, Varadraj Gurupur, …
Hardcover
R5,116
Discovery Miles 51 160
Systems, Decision and Control in Energy…
Artur Zaporozhets, Volodymyr Artemchuk
Hardcover
R4,405
Discovery Miles 44 050
Analysis and Mathematical Models of…
Qiang Yu, Xuesong Wang, …
Hardcover
R2,885
Discovery Miles 28 850
Computational Diffusion MRI - MICCAI…
Elisenda Bonet-Carne, Jana Hutter, …
Hardcover
R4,359
Discovery Miles 43 590
|