![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Computer software packages > Other software packages
Progressively more and more attention has been paid to how location affects health outcomes. The area of disease mapping focusses on these problems, and the Bayesian paradigm has a major role to play in the understanding of the complex interplay of context and individual predisposition in such studies of disease. Using R for Bayesian Spatial and Spatio-Temporal Health Modeling provides a major resource for those interested in applying Bayesian methodology in small area health data studies. Features: Review of R graphics relevant to spatial health data Overview of Bayesian methods and Bayesian hierarchical modeling as applied to spatial data Bayesian Computation and goodness-of-fit Review of basic Bayesian disease mapping models Spatio-temporal modeling with MCMC and INLA Special topics include multivariate models, survival analysis, missing data, measurement error, variable selection, individual event modeling, and infectious disease modeling Software for fitting models based on BRugs, Nimble, CARBayes and INLA Provides code relevant to fitting all examples throughout the book at a supplementary website The book fills a void in the literature and available software, providing a crucial link for students and professionals alike to engage in the analysis of spatial and spatio-temporal health data from a Bayesian perspective using R. The book emphasizes the use of MCMC via Nimble, BRugs, and CARBAyes, but also includes INLA for comparative purposes. In addition, a wide range of packages useful in the analysis of geo-referenced spatial data are employed and code is provided. It will likely become a key reference for researchers and students from biostatistics, epidemiology, public health, and environmental science.
This book provides hands-on guidance for researchers and practitioners in criminal justice and criminology to perform statistical analyses and data visualization in the free and open-source software R. It offers a step-by-step guide for beginners to become familiar with the RStudio platform and tidyverse set of packages. This volume will help users master the fundamentals of the R programming language, providing tutorials in each chapter that lay out research questions and hypotheses centering around a real criminal justice dataset, such as data from the National Survey on Drug Use and Health, National Crime Victimization Survey, Youth Risk Behavior Surveillance System, The Monitoring the Future Study, and The National Youth Survey. Users will also learn how to manipulate common sources of agency data, such as calls-for-service (CFS) data. The end of each chapter includes exercises that reinforce the R tutorial examples, designed to help master the software as well as to provide practice on statistical concepts, data analysis, and interpretation of results. The text can be used as a stand-alone guide to learning R or it can be used as a companion guide to an introductory statistics textbook, such as Basic Statistics in Criminal Justice (2020).
Physically Unclonable Functions (PUFs) translate unavoidable variations in certain parameters of materials, waves, or devices into random and unique signals. They have found many applications in the Internet of Things (IoT), authentication systems, FPGA industry, several other areas in communications and related technologies, and many commercial products. Statistical Trend Analysis of Physically Unclonable Functions first presents a review on cryptographic hardware and hardware-assisted cryptography. The review highlights PUF as a mega trend in research on cryptographic hardware design. Afterwards, the authors present a combined survey and research work on PUFs using a systematic approach. As part of the survey aspect, a state-of-the-art analysis is presented as well as a taxonomy on PUFs, a life cycle, and an established ecosystem for the technology. In another part of the survey, the evolutionary history of PUFs is examined, and strategies for further research in this area are suggested. In the research side, this book presents a novel approach for trend analysis that can be applied to any technology or research area. In this method, a text mining tool is used which extracts 1020 keywords from the titles of the sample papers. Then, a classifying tool classifies the keywords into 295 meaningful research topics. The popularity of each topic is then numerically measured and analyzed over the course of time through a statistical analysis on the number of research papers related to the topic as well as the number of their citations. The authors identify the most popular topics in four different domains; over the history of PUFs, during the recent years, in top conferences, and in top journals. The results are used to present an evolution study as well as a trend analysis and develop a roadmap for future research in this area. This method gives an automatic popularity-based statistical trend analysis which eliminates the need for passing personal judgments about the direction of trends, and provides concrete evidence to the future direction of research on PUFs. Another advantage of this method is the possibility of studying a whole lot of existing research works (more than 700 in this book). This book will appeal to researchers in text mining, cryptography, hardware security, and IoT.
Review of the First Edition: The authors strive to reduce theory to a minimum, which makes it a self-learning text that is comprehensible for biologists, physicians, etc. who lack an advanced mathematics background. Unlike in many other textbooks, R is not introduced with meaningless toy examples; instead the reader is taken by the hand and shown around some analyses, graphics, and simulations directly relating to meta-analysis... A useful hands-on guide for practitioners who want to familiarize themselves with the fundamentals of meta-analysis and get started without having to plough through theorems and proofs. -Journal of Applied Statistics Statistical Meta-Analysis with R and Stata, Second Edition provides a thorough presentation of statistical meta-analyses (MA) with step-by-step implementations using R/Stata. The authors develop analysis step by step using appropriate R/Stata functions, which enables readers to gain an understanding of meta-analysis methods and R/Stata implementation so that they can use these two popular software packages to analyze their own meta-data. Each chapter gives examples of real studies compiled from the literature. After presenting the data and necessary background for understanding the applications, various methods for analyzing meta-data are introduced. The authors then develop analysis code using the appropriate R/Stata packages and functions. What's New in the Second Edition: Adds Stata programs along with the R programs for meta-analysis Updates all the statistical meta-analyses with R/Stata programs Covers fixed-effects and random-effects MA, meta-regression, MA with rare-event, and MA-IPD vs MA-SS Adds five new chapters on multivariate MA, publication bias, missing data in MA, MA in evaluating diagnostic accuracy, and network MA Suitable as a graduate-level text for a meta-data analysis course, the book is also a valuable reference for practitioners and biostatisticians (even those with little or no experience in using R or Stata) in public health, medical research, governmental agencies, and the pharmaceutical industry.
This version replaces R with Python to make it accessible to a greater number of users outside of statistics including those from Machine Learning. A reader coming to this book from an ML background will learn new statistical perspectives on learning from data. Topics include Model Selection, Shrinkage, Experiments with Blocks and Missing Data. Includes an Appendix on Python for beginners.
Simple English format Foundation sections on inference and evidence, and data integration in project management Exploration of R usage through a narrative examining a generic integrated conservation and development project A final section on R for reproducible workflow Accompanied by an R package
This introductory textbook presents research methods and data analysis tools in non-technical language. It explains the research process and the basics of qualitative and quantitative data analysis, including procedures and methods, analysis, interpretation, and applications using hands-on data examples in QDA Miner Lite and IBM SPSS Statistics software. The book is divided into four parts that address study and research design; data collection, qualitative methods and surveys; statistical methods, including hypothesis testing, regression, cluster and factor analysis; and reporting. The intended audience is business and social science students learning scientific research methods, however, given its business context, the book will be equally useful for decision-makers in businesses and organizations.
Praise for the First Edition "The main strength of this book is that it provides a unified framework of graphical tools for data analysis, especially for univariate and low-dimensional multivariate data. In addition, it is clearly written in plain language and the inclusion of R code is particularly useful to assist readers' understanding of the graphical techniques discussed in the book. ... It not only summarises graphical techniques, but it also serves as a practical reference for researchers and graduate students with an interest in data display." -Han Lin Shang, Journal of Applied Statistics Graphics for Statistics and Data Analysis with R, Second Edition, presents the basic principles of graphical design and applies these principles to engaging examples using the graphics and lattice packages in R. It offers a wide array of modern graphical displays for data visualization and representation. Added in the second edition are coverage of the ggplot2 graphics package, material on human visualization and color rendering in R, on screen, and in print. Features Emphasizes the fundamentals of statistical graphics and best practice guidelines for producing and choosing among graphical displays in R Presents technical details on topics such as: the estimation of quantiles, nonparametric and parametric density estimation; diagnostic plots for the simple linear regression model; polynomial regression, splines, and locally weighted polynomial regression for producing a smooth curve; Trellis graphics for multivariate data Provides downloadable R code and data for figures at www.graphicsforstatistics.com Kevin J. Keen is a Professor of Mathematics and Statistics at the University of Northern British Columbia (Prince George, Canada) and an Accredited Professional StatisticianTM by the Statistical Society of Canada and the American Statistical Association.
Microsoft Project 2013 is a powerful software tool, and like all tools it requires knowledge and skill to be used to its maximum potential. This fully revised new edition provides users with everything they will need to more easily and effectively manage projects to a successful conclusion. Designed for the busy, practicing project manager, Dynamic Scheduling With Microsoft Project 2013 will help you get up to speed quickly with the new and enhanced features of Project 2013 (including Project Pro for Office 365) and enable you to create effective schedules using best practices, tips & tricks, and step-by-step instruction. Through the use of helpful screenshots, hands-on exercises, illustrations, and review questions, this guide instructs you on how to build dynamic schedules that will allow you to explore what-if scenarios and dramatically decrease the time you spend making static schedule changes. ""A must read, reread, and use daily for all project managers"" is what PMI's Project Management Journal had to say about previous editions. This updated version is even better!
Mathematica by Example, Sixth Edition is an essential resource for the Mathematica user, providing step-by-step instructions on achieving results from this powerful software tool. The book fully accounts for the changes to functionality and visualization capabilities and accomodates the full array of new extensions in the types of data and problems that Mathematica can immediately handle, including cloud services and systems, geographic and geometric computation, dynamic visualization, interactive applications and other improvements. It is an ideal text for scientific students, researchers, and aspiring programmers seeking further understanding of Mathematica. Written by seasoned practitioners with a view to practical implementation and problem-solving, the book's pedagogy is delivered clearly and without jargon using representative biological, physical and engineering problems. Code is provided on an ancillary website to support the use of Mathematica across diverse applications and subject areas.
Simple English format Foundation sections on inference and evidence, and data integration in project management Exploration of R usage through a narrative examining a generic integrated conservation and development project A final section on R for reproducible workflow Accompanied by an R package
Big Data: A Tutorial-Based Approach explores the tools and techniques used to bring about the marriage of structured and unstructured data. It focuses on Hadoop Distributed Storage and MapReduce Processing by implementing (i) Tools and Techniques of Hadoop Eco System, (ii) Hadoop Distributed File System Infrastructure, and (iii) efficient MapReduce processing. The book includes Use Cases and Tutorials to provide an integrated approach that answers the 'What', 'How', and 'Why' of Big Data. Features Identifies the primary drivers of Big Data Walks readers through the theory, methods and technology of Big Data Explains how to handle the 4 V's of Big Data in order to extract value for better business decision making Shows how and why data connectors are critical and necessary for Agile text analytics Includes in-depth tutorials to perform necessary set-ups, installation, configuration and execution of important tasks Explains the command line as well as GUI interface to a powerful data exchange tool between Hadoop and legacy r-dbms databases
Applied Numerical Methods for Chemical Engineers emphasizes the derivation of a variety of numerical methods and their application to the solution of engineering problems, with special attention to problems in the chemical engineering field. These algorithms encompass linear and nonlinear algebraic equations, eigenvalue problems, finite difference methods, interpolation, differentiation and integration, ordinary differential equations, boundary value problems, partial differential equations, and linear and nonlinear regression analysis. MATLAB is adopted as the calculation environment throughout the book because of its ability to perform all the calculations in matrix form, its large library of built-in functions, its strong structural language, and its rich graphical visualization tools. Through this book, students and other users will learn about the basic features, advantages and disadvantages of various numerical methods, learn and practice many useful m-files developed for different numerical methods in addition to the MATLAB built-in solvers, develop and set up mathematical models for problems commonly encountered in chemical engineering, and solve chemical engineering related problems through examples and after-chapter problems with MATLAB by creating application m-files.
Praise for the first edition: [This book] reflects the extensive experience and significant contributions of the author to non-linear and non-Gaussian modeling. ... [It] is a valuable book, especially with its broad and accessible introduction of models in the state-space framework. -Statistics in Medicine What distinguishes this book from comparable introductory texts is the use of state-space modeling. Along with this come a number of valuable tools for recursive filtering and smoothing, including the Kalman filter, as well as non-Gaussian and sequential Monte Carlo filters. -MAA Reviews Introduction to Time Series Modeling with Applications in R, Second Edition covers numerous stationary and nonstationary time series models and tools for estimating and utilizing them. The goal of this book is to enable readers to build their own models to understand, predict and master time series. The second edition makes it possible for readers to reproduce examples in this book by using the freely available R package TSSS to perform computations for their own real-world time series problems. This book employs the state-space model as a generic tool for time series modeling and presents the Kalman filter, the non-Gaussian filter and the particle filter as convenient tools for recursive estimation for state-space models. Further, it also takes a unified approach based on the entropy maximization principle and employs various methods of parameter estimation and model selection, including the least squares method, the maximum likelihood method, recursive estimation for state-space models and model selection by AIC. Along with the standard stationary time series models, such as the AR and ARMA models, the book also introduces nonstationary time series models such as the locally stationary AR model, the trend model, the seasonal adjustment model, the time-varying coefficient AR model and nonlinear non-Gaussian state-space models. About the Author: Genshiro Kitagawa is a project professor at the University of Tokyo, the former Director-General of the Institute of Statistical Mathematics, and the former President of the Research Organization of Information and Systems.
Modeling spatial and spatio-temporal continuous processes is an important and challenging problem in spatial statistics. Advanced Spatial Modeling with Stochastic Partial Differential Equations Using R and INLA describes in detail the stochastic partial differential equations (SPDE) approach for modeling continuous spatial processes with a Matern covariance, which has been implemented using the integrated nested Laplace approximation (INLA) in the R-INLA package. Key concepts about modeling spatial processes and the SPDE approach are explained with examples using simulated data and real applications. This book has been authored by leading experts in spatial statistics, including the main developers of the INLA and SPDE methodologies and the R-INLA package. It also includes a wide range of applications: * Spatial and spatio-temporal models for continuous outcomes * Analysis of spatial and spatio-temporal point patterns * Coregionalization spatial and spatio-temporal models * Measurement error spatial models * Modeling preferential sampling * Spatial and spatio-temporal models with physical barriers * Survival analysis with spatial effects * Dynamic space-time regression * Spatial and spatio-temporal models for extremes * Hurdle models with spatial effects * Penalized Complexity priors for spatial models All the examples in the book are fully reproducible. Further information about this book, as well as the R code and datasets used, is available from the book website at http://www.r-inla.org/spde-book. The tools described in this book will be useful to researchers in many fields such as biostatistics, spatial statistics, environmental sciences, epidemiology, ecology and others. Graduate and Ph.D. students will also find this book and associated files a valuable resource to learn INLA and the SPDE approach for spatial modeling.
This unique book addresses the statistical modelling and analysis of microbiome data using cutting-edge R software. It includes real-world data from the authors' research and from the public domain, and discusses the implementation of R for data analysis step by step. The data and R computer programs are publicly available, allowing readers to replicate the model development and data analysis presented in each chapter, so that these new methods can be readily applied in their own research. The book also discusses recent developments in statistical modelling and data analysis in microbiome research, as well as the latest advances in next-generation sequencing and big data in methodological development and applications. This timely book will greatly benefit all readers involved in microbiome, ecology and microarray data analyses, as well as other fields of research.
A hands-on guide to using Excel in the business context First published in 2012, Using Excel for Business and Financial Modelling contains step-by-step instructions of how to solve common business problems using financial models, including downloadable Excel templates, a list of shortcuts and tons of practical tips and techniques you can apply straight away. Whilst there are many hundreds of tools, features and functions in Excel, this book focuses on the topics most relevant to finance professionals. It covers these features in detail from a practical perspective, but also puts them in context by applying them to practical examples in the real world. Learn to create financial models to help make business decisions whilst applying modelling best practice methodology, tools and techniques. - Provides the perfect mix of practice and theory - Helps you become a DIY Excel modelling specialist - Includes updates for Excel 2019/365 and Excel for Mac - May be used as an accompaniment to the author's online and face-to-face training courses Many people are often overwhelmed by the hundreds of tools in Excel, and this book gives clarity to the ones you need to know in order to perform your job more efficiently. This book also demystifies the technical, design, logic and financial skills you need for business and financial modelling.
R Visualizations: Derive Meaning from Data focuses on one of the two major topics of data analytics: data visualization, a.k.a., computer graphics. In the book, major R systems for visualization are discussed, organized by topic and not by system. Anyone doing data analysis will be shown how to use R to generate any of the basic visualizations with the R visualization systems. Further, this book introduces the author's lessR system, which always can accomplish a visualization with less coding than the use of other systems, sometimes dramatically so, and also provides accompanying statistical analyses. Key Features Presents thorough coverage of the leading R visualization system, ggplot2. Gives specific guidance on using base R graphics to attain visualizations of the same quality as those provided by ggplot2. Shows how to create a wide range of data visualizations: distributions of categorical and continuous variables, many types of scatterplots including with a third variable, time series, and maps. Inclusion of the various approaches to R graphics organized by topic instead of by system. Presents the recent work on interactive visualization in R. David W. Gerbing received his PhD from Michigan State University in 1979 in quantitative analysis, and currently is a professor of quantitative analysis in the School of Business at Portland State University. He has published extensively in the social and behavioral sciences with a focus on quantitative methods. His lessR package has been in development since 2009.
Technological improvements continue to push back the frontier of processor speed in modern computers. Unfortunately, the computational intensity demanded by modern research problems grows even faster. Parallel computing has emerged as the most successful bridge to this computational gap, and many popular solutions have emerged based on its concepts, such as grid computing and massively parallel supercomputers. The Handbook of Parallel Computing and Statistics systematically applies the principles of parallel computing for solving increasingly complex problems in statistics research. This unique reference weaves together the principles and theoretical models of parallel computing with the design, analysis, and application of algorithms for solving statistical problems. After a brief introduction to parallel computing, the book explores the architecture, programming, and computational aspects of parallel processing. Focus then turns to optimization methods followed by statistical applications. These applications include algorithms for predictive modeling, adaptive design, real-time estimation of higher-order moments and cumulants, data mining, econometrics, and Bayesian computation. Expert contributors summarize recent results and explore new directions in these areas. Its intricate combination of theory and practical applications makes the Handbook of Parallel Computing and Statistics an ideal companion for helping solve the abundance of computation-intensive statistical problems arising in a variety of fields.
This volume features original contributions and invited review articles on mathematical statistics, statistical simulation and experimental design. The selected peer-reviewed contributions originate from the 8th International Workshop on Simulation held in Vienna in 2015. The book is intended for mathematical statisticians, Ph.D. students and statisticians working in medicine, engineering, pharmacy, psychology, agriculture and other related fields. The International Workshops on Simulation are devoted to statistical techniques in stochastic simulation, data collection, design of scientific experiments and studies representing broad areas of interest. The first 6 workshops took place in St. Petersburg, Russia, in 1994 - 2009 and the 7th workshop was held in Rimini, Italy, in 2013.
This workbook and solutions manual is intended for advanced undergraduate or beginning graduate students as a supplement to a traditional course in numerical mathematics and as preparation for independent research involving numerical mathematics. The solutions manual provides complete MATLAB code and numerical results for each of the exercises in the workbook and will be especially useful for those students without previous MATLAB programming experience. It is also valuable for classroom instructors to help pinpoint the author's intent in each exercise and to provide a model for graders. Upon completion of this material, students will have a working knowledge of MATLAB programming, they will have themselves programmed algorithms encountered in classwork and textbooks, and they will know how to check and verify their own programs against hand calculations and by reference to theoretical results, special polynomial solutions and other specialized solutions. No previous programming experience with MATLAB is necessary.
In social science outstanding results are yielded by advanced simulation methods, based on state of the art software technologies and an appropriate combination of qualitative and quantitative methods. This book presents examples of successful applications of modelling and computing in social science: business and logistic process simulation and optimization, deeper knowledge extractions from big data, better understanding and predicting of social behaviour and modelling health and environment changes.
The richly illustrated Interactive Web-Based Data Visualization with R, plotly, and shiny focuses on the process of programming interactive web graphics for multidimensional data analysis. It is written for the data analyst who wants to leverage the capabilities of interactive web graphics without having to learn web programming. Through many R code examples, you will learn how to tap the extensive functionality of these tools to enhance the presentation and exploration of data. By mastering these concepts and tools, you will impress your colleagues with your ability to quickly generate more informative, engaging, and reproducible interactive graphics using free and open source software that you can share over email, export to pdf, and more. Key Features: Convert static ggplot2 graphics to an interactive web-based form Link, animate, and arrange multiple plots in standalone HTML from R Embed, modify, and respond to plotly graphics in a shiny app Learn best practices for visualizing continuous, discrete, and multivariate data Learn numerous ways to visualize geo-spatial data This book makes heavy use of plotly for graphical rendering, but you will also learn about other R packages that support different phases of a data science workflow, such as tidyr, dplyr, and tidyverse. Along the way, you will gain insight into best practices for visualization of high-dimensional data, statistical graphics, and graphical perception. The printed book is complemented by an interactive website where readers can view movies demonstrating the examples and interact with graphics. |
![]() ![]() You may like...
White Fragility - Why It's So Hard For…
Robin DiAngelo
Paperback
![]()
Recent Advances in Psychology and Aging…
P. Costa, I. C. Siegler
Hardcover
R3,714
Discovery Miles 37 140
|