![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer software packages > Other software packages
Progressively more and more attention has been paid to how location affects health outcomes. The area of disease mapping focusses on these problems, and the Bayesian paradigm has a major role to play in the understanding of the complex interplay of context and individual predisposition in such studies of disease. Using R for Bayesian Spatial and Spatio-Temporal Health Modeling provides a major resource for those interested in applying Bayesian methodology in small area health data studies. Features: Review of R graphics relevant to spatial health data Overview of Bayesian methods and Bayesian hierarchical modeling as applied to spatial data Bayesian Computation and goodness-of-fit Review of basic Bayesian disease mapping models Spatio-temporal modeling with MCMC and INLA Special topics include multivariate models, survival analysis, missing data, measurement error, variable selection, individual event modeling, and infectious disease modeling Software for fitting models based on BRugs, Nimble, CARBayes and INLA Provides code relevant to fitting all examples throughout the book at a supplementary website The book fills a void in the literature and available software, providing a crucial link for students and professionals alike to engage in the analysis of spatial and spatio-temporal health data from a Bayesian perspective using R. The book emphasizes the use of MCMC via Nimble, BRugs, and CARBAyes, but also includes INLA for comparative purposes. In addition, a wide range of packages useful in the analysis of geo-referenced spatial data are employed and code is provided. It will likely become a key reference for researchers and students from biostatistics, epidemiology, public health, and environmental science.
This book introduces the basic methodologies for successful data analytics. Matrix optimization and approximation are explained in detail and extensively applied to dimensionality reduction by principal component analysis and multidimensional scaling. Diffusion maps and spectral clustering are derived as powerful tools. The methodological overlap between data science and machine learning is emphasized by demonstrating how data science is used for classification as well as supervised and unsupervised learning.
Review of the First Edition: The authors strive to reduce theory to a minimum, which makes it a self-learning text that is comprehensible for biologists, physicians, etc. who lack an advanced mathematics background. Unlike in many other textbooks, R is not introduced with meaningless toy examples; instead the reader is taken by the hand and shown around some analyses, graphics, and simulations directly relating to meta-analysis... A useful hands-on guide for practitioners who want to familiarize themselves with the fundamentals of meta-analysis and get started without having to plough through theorems and proofs. -Journal of Applied Statistics Statistical Meta-Analysis with R and Stata, Second Edition provides a thorough presentation of statistical meta-analyses (MA) with step-by-step implementations using R/Stata. The authors develop analysis step by step using appropriate R/Stata functions, which enables readers to gain an understanding of meta-analysis methods and R/Stata implementation so that they can use these two popular software packages to analyze their own meta-data. Each chapter gives examples of real studies compiled from the literature. After presenting the data and necessary background for understanding the applications, various methods for analyzing meta-data are introduced. The authors then develop analysis code using the appropriate R/Stata packages and functions. What's New in the Second Edition: Adds Stata programs along with the R programs for meta-analysis Updates all the statistical meta-analyses with R/Stata programs Covers fixed-effects and random-effects MA, meta-regression, MA with rare-event, and MA-IPD vs MA-SS Adds five new chapters on multivariate MA, publication bias, missing data in MA, MA in evaluating diagnostic accuracy, and network MA Suitable as a graduate-level text for a meta-data analysis course, the book is also a valuable reference for practitioners and biostatisticians (even those with little or no experience in using R or Stata) in public health, medical research, governmental agencies, and the pharmaceutical industry.
Physically Unclonable Functions (PUFs) translate unavoidable variations in certain parameters of materials, waves, or devices into random and unique signals. They have found many applications in the Internet of Things (IoT), authentication systems, FPGA industry, several other areas in communications and related technologies, and many commercial products. Statistical Trend Analysis of Physically Unclonable Functions first presents a review on cryptographic hardware and hardware-assisted cryptography. The review highlights PUF as a mega trend in research on cryptographic hardware design. Afterwards, the authors present a combined survey and research work on PUFs using a systematic approach. As part of the survey aspect, a state-of-the-art analysis is presented as well as a taxonomy on PUFs, a life cycle, and an established ecosystem for the technology. In another part of the survey, the evolutionary history of PUFs is examined, and strategies for further research in this area are suggested. In the research side, this book presents a novel approach for trend analysis that can be applied to any technology or research area. In this method, a text mining tool is used which extracts 1020 keywords from the titles of the sample papers. Then, a classifying tool classifies the keywords into 295 meaningful research topics. The popularity of each topic is then numerically measured and analyzed over the course of time through a statistical analysis on the number of research papers related to the topic as well as the number of their citations. The authors identify the most popular topics in four different domains; over the history of PUFs, during the recent years, in top conferences, and in top journals. The results are used to present an evolution study as well as a trend analysis and develop a roadmap for future research in this area. This method gives an automatic popularity-based statistical trend analysis which eliminates the need for passing personal judgments about the direction of trends, and provides concrete evidence to the future direction of research on PUFs. Another advantage of this method is the possibility of studying a whole lot of existing research works (more than 700 in this book). This book will appeal to researchers in text mining, cryptography, hardware security, and IoT.
This book illustrates the role of software architecture and its application in business. The author describes enterprise architecture along with business architecture to show the role of software architecture in both areas. The place of software architecture in business is outlined from many perspectives in this context. The book outlines quality attributes and how managers can use software architecture to build high quality products. Topics include business software architecture, dealing with qualities, achieving quality attributes, managing business qualities, software product line, Internet of Things (IOT), and Service Oriented Business Architecture. The book is intended to benefit students, researchers, software architects, and business architects. Provides quick and easy access to all the important aspects of software architecture in business; Highlights a wide variety of concepts of software architecture in a straightforward manner, for students, practitioners, or architects; Presents different applications of software architecture in business.
Simple English format Foundation sections on inference and evidence, and data integration in project management Exploration of R usage through a narrative examining a generic integrated conservation and development project A final section on R for reproducible workflow Accompanied by an R package
Simple English format Foundation sections on inference and evidence, and data integration in project management Exploration of R usage through a narrative examining a generic integrated conservation and development project A final section on R for reproducible workflow Accompanied by an R package
Praise for the First Edition "The main strength of this book is that it provides a unified framework of graphical tools for data analysis, especially for univariate and low-dimensional multivariate data. In addition, it is clearly written in plain language and the inclusion of R code is particularly useful to assist readers' understanding of the graphical techniques discussed in the book. ... It not only summarises graphical techniques, but it also serves as a practical reference for researchers and graduate students with an interest in data display." -Han Lin Shang, Journal of Applied Statistics Graphics for Statistics and Data Analysis with R, Second Edition, presents the basic principles of graphical design and applies these principles to engaging examples using the graphics and lattice packages in R. It offers a wide array of modern graphical displays for data visualization and representation. Added in the second edition are coverage of the ggplot2 graphics package, material on human visualization and color rendering in R, on screen, and in print. Features Emphasizes the fundamentals of statistical graphics and best practice guidelines for producing and choosing among graphical displays in R Presents technical details on topics such as: the estimation of quantiles, nonparametric and parametric density estimation; diagnostic plots for the simple linear regression model; polynomial regression, splines, and locally weighted polynomial regression for producing a smooth curve; Trellis graphics for multivariate data Provides downloadable R code and data for figures at www.graphicsforstatistics.com Kevin J. Keen is a Professor of Mathematics and Statistics at the University of Northern British Columbia (Prince George, Canada) and an Accredited Professional StatisticianTM by the Statistical Society of Canada and the American Statistical Association.
This version replaces R with Python to make it accessible to a greater number of users outside of statistics including those from Machine Learning. A reader coming to this book from an ML background will learn new statistical perspectives on learning from data. Topics include Model Selection, Shrinkage, Experiments with Blocks and Missing Data. Includes an Appendix on Python for beginners.
Signals and Systems: A Primer with MATLAB (R) provides clear, interesting, and easy-to-understand coverage of continuous-time and discrete-time signals and systems. Each chapter opens with a historical profile or career talk, followed by an introduction that states the chapter objectives and links the chapter to the previous ones. All principles are presented in a lucid, logical, step-by-step approach. As much as possible, the authors avoid wordiness and detail overload that could hide concepts and impede understanding. In recognition of the requirements by the Accreditation Board for Engineering and Technology (ABET) on integrating computer tools, the use of MATLAB (R) is encouraged in a student-friendly manner. MATLAB is introduced in Appendix B and applied gradually throughout the book. Each illustrative example is immediately followed by a practice problem along with its answer. Students can follow the example step by step to solve the practice problem without flipping pages or looking at the end of the book for answers. These practice problems test students' comprehension and reinforce key concepts before moving on to the next section. Toward the end of each chapter, the authors discuss some application aspects of the concepts covered in the chapter. The material covered in the chapter is applied to at least one or two practical problems or devices. This helps students see how the concepts are applied to real-life situations. In addition, thoroughly worked examples are given liberally at the end of every section. These examples give students a solid grasp of the solutions as well as the confidence to solve similar problems themselves. Some of the problems are solved in two or three ways to facilitate a deeper understanding and comparison of different approaches. Ten review questions in the form of multiple-choice objective items are provided at the end of each chapter with answers. The review questions are intended to cover the "little tricks" that the examples and end-of-chapter problems may not cover. They serve as a self-test device and help students determine chapter mastery. Each chapter also ends with a summary of key points and formulas. Designed for a three-hour semester course on signals and systems, Signals and Systems: A Primer with MATLAB (R) is intended as a textbook for junior-level undergraduate students in electrical and computer engineering. The prerequisites for a course based on this book are knowledge of standard mathematics (including calculus and differential equations) and electric circuit analysis.
Praise for the first edition: [This book] reflects the extensive experience and significant contributions of the author to non-linear and non-Gaussian modeling. ... [It] is a valuable book, especially with its broad and accessible introduction of models in the state-space framework. -Statistics in Medicine What distinguishes this book from comparable introductory texts is the use of state-space modeling. Along with this come a number of valuable tools for recursive filtering and smoothing, including the Kalman filter, as well as non-Gaussian and sequential Monte Carlo filters. -MAA Reviews Introduction to Time Series Modeling with Applications in R, Second Edition covers numerous stationary and nonstationary time series models and tools for estimating and utilizing them. The goal of this book is to enable readers to build their own models to understand, predict and master time series. The second edition makes it possible for readers to reproduce examples in this book by using the freely available R package TSSS to perform computations for their own real-world time series problems. This book employs the state-space model as a generic tool for time series modeling and presents the Kalman filter, the non-Gaussian filter and the particle filter as convenient tools for recursive estimation for state-space models. Further, it also takes a unified approach based on the entropy maximization principle and employs various methods of parameter estimation and model selection, including the least squares method, the maximum likelihood method, recursive estimation for state-space models and model selection by AIC. Along with the standard stationary time series models, such as the AR and ARMA models, the book also introduces nonstationary time series models such as the locally stationary AR model, the trend model, the seasonal adjustment model, the time-varying coefficient AR model and nonlinear non-Gaussian state-space models. About the Author: Genshiro Kitagawa is a project professor at the University of Tokyo, the former Director-General of the Institute of Statistical Mathematics, and the former President of the Research Organization of Information and Systems.
Big Data: A Tutorial-Based Approach explores the tools and techniques used to bring about the marriage of structured and unstructured data. It focuses on Hadoop Distributed Storage and MapReduce Processing by implementing (i) Tools and Techniques of Hadoop Eco System, (ii) Hadoop Distributed File System Infrastructure, and (iii) efficient MapReduce processing. The book includes Use Cases and Tutorials to provide an integrated approach that answers the 'What', 'How', and 'Why' of Big Data. Features Identifies the primary drivers of Big Data Walks readers through the theory, methods and technology of Big Data Explains how to handle the 4 V's of Big Data in order to extract value for better business decision making Shows how and why data connectors are critical and necessary for Agile text analytics Includes in-depth tutorials to perform necessary set-ups, installation, configuration and execution of important tasks Explains the command line as well as GUI interface to a powerful data exchange tool between Hadoop and legacy r-dbms databases
This volume features original contributions and invited review articles on mathematical statistics, statistical simulation and experimental design. The selected peer-reviewed contributions originate from the 8th International Workshop on Simulation held in Vienna in 2015. The book is intended for mathematical statisticians, Ph.D. students and statisticians working in medicine, engineering, pharmacy, psychology, agriculture and other related fields. The International Workshops on Simulation are devoted to statistical techniques in stochastic simulation, data collection, design of scientific experiments and studies representing broad areas of interest. The first 6 workshops took place in St. Petersburg, Russia, in 1994 - 2009 and the 7th workshop was held in Rimini, Italy, in 2013.
Modeling spatial and spatio-temporal continuous processes is an important and challenging problem in spatial statistics. Advanced Spatial Modeling with Stochastic Partial Differential Equations Using R and INLA describes in detail the stochastic partial differential equations (SPDE) approach for modeling continuous spatial processes with a Matern covariance, which has been implemented using the integrated nested Laplace approximation (INLA) in the R-INLA package. Key concepts about modeling spatial processes and the SPDE approach are explained with examples using simulated data and real applications. This book has been authored by leading experts in spatial statistics, including the main developers of the INLA and SPDE methodologies and the R-INLA package. It also includes a wide range of applications: * Spatial and spatio-temporal models for continuous outcomes * Analysis of spatial and spatio-temporal point patterns * Coregionalization spatial and spatio-temporal models * Measurement error spatial models * Modeling preferential sampling * Spatial and spatio-temporal models with physical barriers * Survival analysis with spatial effects * Dynamic space-time regression * Spatial and spatio-temporal models for extremes * Hurdle models with spatial effects * Penalized Complexity priors for spatial models All the examples in the book are fully reproducible. Further information about this book, as well as the R code and datasets used, is available from the book website at http://www.r-inla.org/spde-book. The tools described in this book will be useful to researchers in many fields such as biostatistics, spatial statistics, environmental sciences, epidemiology, ecology and others. Graduate and Ph.D. students will also find this book and associated files a valuable resource to learn INLA and the SPDE approach for spatial modeling.
Financial, Macro and Micro Econometrics Using R, Volume 42, provides state-of-the-art information on important topics in econometrics, including multivariate GARCH, stochastic frontiers, fractional responses, specification testing and model selection, exogeneity testing, causal analysis and forecasting, GMM models, asset bubbles and crises, corporate investments, classification, forecasting, nonstandard problems, cointegration, financial market jumps and co-jumps, among other topics.
This unique book addresses the statistical modelling and analysis of microbiome data using cutting-edge R software. It includes real-world data from the authors' research and from the public domain, and discusses the implementation of R for data analysis step by step. The data and R computer programs are publicly available, allowing readers to replicate the model development and data analysis presented in each chapter, so that these new methods can be readily applied in their own research. The book also discusses recent developments in statistical modelling and data analysis in microbiome research, as well as the latest advances in next-generation sequencing and big data in methodological development and applications. This timely book will greatly benefit all readers involved in microbiome, ecology and microarray data analyses, as well as other fields of research.
R Visualizations: Derive Meaning from Data focuses on one of the two major topics of data analytics: data visualization, a.k.a., computer graphics. In the book, major R systems for visualization are discussed, organized by topic and not by system. Anyone doing data analysis will be shown how to use R to generate any of the basic visualizations with the R visualization systems. Further, this book introduces the author's lessR system, which always can accomplish a visualization with less coding than the use of other systems, sometimes dramatically so, and also provides accompanying statistical analyses. Key Features Presents thorough coverage of the leading R visualization system, ggplot2. Gives specific guidance on using base R graphics to attain visualizations of the same quality as those provided by ggplot2. Shows how to create a wide range of data visualizations: distributions of categorical and continuous variables, many types of scatterplots including with a third variable, time series, and maps. Inclusion of the various approaches to R graphics organized by topic instead of by system. Presents the recent work on interactive visualization in R. David W. Gerbing received his PhD from Michigan State University in 1979 in quantitative analysis, and currently is a professor of quantitative analysis in the School of Business at Portland State University. He has published extensively in the social and behavioral sciences with a focus on quantitative methods. His lessR package has been in development since 2009.
Technological improvements continue to push back the frontier of processor speed in modern computers. Unfortunately, the computational intensity demanded by modern research problems grows even faster. Parallel computing has emerged as the most successful bridge to this computational gap, and many popular solutions have emerged based on its concepts, such as grid computing and massively parallel supercomputers. The Handbook of Parallel Computing and Statistics systematically applies the principles of parallel computing for solving increasingly complex problems in statistics research. This unique reference weaves together the principles and theoretical models of parallel computing with the design, analysis, and application of algorithms for solving statistical problems. After a brief introduction to parallel computing, the book explores the architecture, programming, and computational aspects of parallel processing. Focus then turns to optimization methods followed by statistical applications. These applications include algorithms for predictive modeling, adaptive design, real-time estimation of higher-order moments and cumulants, data mining, econometrics, and Bayesian computation. Expert contributors summarize recent results and explore new directions in these areas. Its intricate combination of theory and practical applications makes the Handbook of Parallel Computing and Statistics an ideal companion for helping solve the abundance of computation-intensive statistical problems arising in a variety of fields.
This introductory textbook presents research methods and data analysis tools in non-technical language. It explains the research process and the basics of qualitative and quantitative data analysis, including procedures and methods, analysis, interpretation, and applications using hands-on data examples in QDA Miner Lite and IBM SPSS Statistics software. The book is divided into four parts that address study and research design; data collection, qualitative methods and surveys; statistical methods, including hypothesis testing, regression, cluster and factor analysis; and reporting. The intended audience is business and social science students learning scientific research methods, however, given its business context, the book will be equally useful for decision-makers in businesses and organizations.
"Industrial Statistics with MINITAB" demonstrates the use of MINITAB as a tool for performing statistical analysis in an industrial context. This book covers introductory industrial statistics, exploring the most commonly used techniques alongside those that serve to give an overview of more complex issues. A plethora of examples in MINITAB are featured along with case studies for each of the statistical techniques presented. "Industrial Statistics with MINITAB" Provides comprehensive coverage of user-friendly practical guidance to the essential statistical methods applied in industry.Explores statistical techniques and how they can be used effectively with the help of MINITAB 16.Contains extensive illustrative examples and case studies throughout and assumes no previous statistical knowledge.Emphasises data graphics and visualization, and the most used industrial statistical tools, such as Statistical Process Control and Design of Experiments. Is supported by an accompanying website featuring case studies and the corresponding datasets. Six Sigma Green Belts and Black Belts will find explanations and examples of the most relevant techniques in DMAIC projects. The book can also be used as quick reference enabling the reader to be confident enough to explore other MINITAB capabilities.
This comprehensive text covers the use of SAS for epidemiology and public health research. Developed with students in mind and from their feedback, the text addresses this material in a straightforward manner with a multitude of examples. It is directly applicable to students and researchers in the fields of public health, biostatistics and epidemiology. Through a hands on approach to the use of SAS for a broad number of epidemiologic analyses, readers learn techniques for data entry and cleaning, categorical analysis, ANOVA, and linear regression and much more. Exercises utilizing real-world data sets are featured throughout the book. SAS screen shots demonstrate the steps for successful programming. SAS (Statistical Analysis System) is an integrated system of software products provided by the SAS institute, which is headquartered in California. It provides programmers and statisticians the ability to engage in many sophisticated statistical analyses and data retrieval and mining exercises. SAS is widely used in the fields of epidemiology and public healthresearch, predominately due to its ability to reliably analyze very large administrative data sets, as well as more commonly encountered clinical trial and observational research data. "
The richly illustrated Interactive Web-Based Data Visualization with R, plotly, and shiny focuses on the process of programming interactive web graphics for multidimensional data analysis. It is written for the data analyst who wants to leverage the capabilities of interactive web graphics without having to learn web programming. Through many R code examples, you will learn how to tap the extensive functionality of these tools to enhance the presentation and exploration of data. By mastering these concepts and tools, you will impress your colleagues with your ability to quickly generate more informative, engaging, and reproducible interactive graphics using free and open source software that you can share over email, export to pdf, and more. Key Features: Convert static ggplot2 graphics to an interactive web-based form Link, animate, and arrange multiple plots in standalone HTML from R Embed, modify, and respond to plotly graphics in a shiny app Learn best practices for visualizing continuous, discrete, and multivariate data Learn numerous ways to visualize geo-spatial data This book makes heavy use of plotly for graphical rendering, but you will also learn about other R packages that support different phases of a data science workflow, such as tidyr, dplyr, and tidyverse. Along the way, you will gain insight into best practices for visualization of high-dimensional data, statistical graphics, and graphical perception. The printed book is complemented by an interactive website where readers can view movies demonstrating the examples and interact with graphics.
This book explores inductive inference using the minimum message length (MML) principle, a Bayesian method which is a realisation of Ockham's Razor based on information theory. Accompanied by a library of software, the book can assist an applications programmer, student or researcher in the fields of data analysis and machine learning to write computer programs based upon this principle. MML inference has been around for 50 years and yet only one highly technical book has been written about the subject. The majority of research in the field has been backed by specialised one-off programs but this book includes a library of general MML-based software, in Java. The Java source code is available under the GNU GPL open-source license. The software library is documented using Javadoc which produces extensive cross referenced HTML manual pages. Every probability distribution and statistical model that is described in the book is implemented and documented in the software library. The library may contain a component that directly solves a reader's inference problem, or contain components that can be put together to solve the problem, or provide a standard interface under which a new component can be written to solve the problem. This book will be of interest to application developers in the fields of machine learning and statistics as well as academics, postdocs, programmers and data scientists. It could also be used by third year or fourth year undergraduate or postgraduate students.
This book systematically addresses the design and analysis of efficient techniques for independent random sampling. Both general-purpose approaches, which can be used to generate samples from arbitrary probability distributions, and tailored techniques, designed to efficiently address common real-world practical problems, are introduced and discussed in detail. In turn, the monograph presents fundamental results and methodologies in the field, elaborating and developing them into the latest techniques. The theory and methods are illustrated with a varied collection of examples, which are discussed in detail in the text and supplemented with ready-to-run computer code. The main problem addressed in the book is how to generate independent random samples from an arbitrary probability distribution with the weakest possible constraints or assumptions in a form suitable for practical implementation. The authors review the fundamental results and methods in the field, address the latest methods, and emphasize the links and interplay between ostensibly diverse techniques. |
You may like...
Database Systems - Design…
Carlos Coronel, Steven Morris
Paperback
14th International Symposium on Process…
Yoshiyuki Yamashita, Manabu Kano
Hardcover
R11,098
Discovery Miles 110 980
Intelligent Edge Computing for Cyber…
D. Jude Hemanth, Bb Gupta, …
Paperback
R2,954
Discovery Miles 29 540
Essential Java for Scientists and…
Brian Hahn, Katherine Malan
Paperback
R1,266
Discovery Miles 12 660
Handbook of HydroInformatics - Volume…
Saeid Eslamian, Faezeh Eslamian
Paperback
R3,507
Discovery Miles 35 070
|