![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer software packages > Other software packages
R Visualizations: Derive Meaning from Data focuses on one of the two major topics of data analytics: data visualization, a.k.a., computer graphics. In the book, major R systems for visualization are discussed, organized by topic and not by system. Anyone doing data analysis will be shown how to use R to generate any of the basic visualizations with the R visualization systems. Further, this book introduces the author's lessR system, which always can accomplish a visualization with less coding than the use of other systems, sometimes dramatically so, and also provides accompanying statistical analyses. Key Features Presents thorough coverage of the leading R visualization system, ggplot2. Gives specific guidance on using base R graphics to attain visualizations of the same quality as those provided by ggplot2. Shows how to create a wide range of data visualizations: distributions of categorical and continuous variables, many types of scatterplots including with a third variable, time series, and maps. Inclusion of the various approaches to R graphics organized by topic instead of by system. Presents the recent work on interactive visualization in R. David W. Gerbing received his PhD from Michigan State University in 1979 in quantitative analysis, and currently is a professor of quantitative analysis in the School of Business at Portland State University. He has published extensively in the social and behavioral sciences with a focus on quantitative methods. His lessR package has been in development since 2009.
Behavior Analysis with Machine Learning Using R introduces machine learning and deep learning concepts and algorithms applied to a diverse set of behavior analysis problems. It focuses on the practical aspects of solving such problems based on data collected from sensors or stored in electronic records. The included examples demonstrate how to perform common data analysis tasks such as: data exploration, visualization, preprocessing, data representation, model training and evaluation. All of this, using the R programming language and real-life behavioral data. Even though the examples focus on behavior analysis tasks, the covered underlying concepts and methods can be applied in any other domain. No prior knowledge in machine learning is assumed. Basic experience with R and basic knowledge in statistics and high school level mathematics are beneficial. Features: Build supervised machine learning models to predict indoor locations based on WiFi signals, recognize physical activities from smartphone sensors and 3D skeleton data, detect hand gestures from accelerometer signals, and so on. Program your own ensemble learning methods and use Multi-View Stacking to fuse signals from heterogeneous data sources. Use unsupervised learning algorithms to discover criminal behavioral patterns. Build deep learning neural networks with TensorFlow and Keras to classify muscle activity from electromyography signals and Convolutional Neural Networks to detect smiles in images. Evaluate the performance of your models in traditional and multi-user settings. Build anomaly detection models such as Isolation Forests and autoencoders to detect abnormal fish behaviors. This book is intended for undergraduate/graduate students and researchers from ubiquitous computing, behavioral ecology, psychology, e-health, and other disciplines who want to learn the basics of machine learning and deep learning and for the more experienced individuals who want to apply machine learning to analyze behavioral data.
This book provides a groundbreaking introduction to the likelihood inference for correlated survival data via the hierarchical (or h-) likelihood in order to obtain the (marginal) likelihood and to address the computational difficulties in inferences and extensions. The approach presented in the book overcomes shortcomings in the traditional likelihood-based methods for clustered survival data such as intractable integration. The text includes technical materials such as derivations and proofs in each chapter, as well as recently developed software programs in R ("frailtyHL"), while the real-world data examples together with an R package, "frailtyHL" in CRAN, provide readers with useful hands-on tools. Reviewing new developments since the introduction of the h-likelihood to survival analysis (methods for interval estimation of the individual frailty and for variable selection of the fixed effects in the general class of frailty models) and guiding future directions, the book is of interest to researchers in medical and genetics fields, graduate students, and PhD (bio) statisticians.
Coherent treatment of a variety of approaches to multiple comparisons Broad coverage of topics, with contributions by internationally leading experts Detailed treatment of applications in medicine and life sciences Suitable for researchers, lecturers / students, and practitioners
Nearly every large corporation and governmental agency is taking a fresh look at their current enterprise-scale business intelligence (BI) and data warehousing implementations at the dawn of the "Big Data Era"...and most see a critical need to revitalize their current capabilities. Whether they find the frustrating and business-impeding continuation of a long-standing "silos of data" problem, or an over-reliance on static production reports at the expense of predictive analytics and other true business intelligence capabilities, or a lack of progress in achieving the long-sought-after enterprise-wide "single version of the truth" - or all of the above - IT Directors, strategists, and architects find that they need to go back to the drawing board and produce a brand new BI/data warehousing roadmap to help move their enterprises from their current state to one where the promises of emerging technologies and a generation's worth of best practices can finally deliver high-impact, architecturally evolvable enterprise-scale business intelligence and data warehousing. Author Alan Simon, whose BI and data warehousing experience dates back to the late 1970s and who has personally delivered or led more than thirty enterprise-wide BI/data warehousing roadmap engagements since the mid-1990s, details a comprehensive step-by-step approach to building a best practices-driven, multi-year roadmap in the quest for architecturally evolvable BI and data warehousing at the enterprise scale. Simon addresses the triad of technology, work processes, and organizational/human factors considerations in a manner that blends the visionary and the pragmatic.
*Systematically introducing major components of SPM process. *Novel hybrid methods (228 hybrids plus numerous variants) of modern statistical methods or machine learning methods with mathematical and/or univariate geostatistical methods. *Novel predictive accuracy-based variable selection techniques for spatial predictive methods. *Predictive accuracy-based parameter/model optimization. *Reproducible examples for SPM of various data types in R.
Data Analytics for the Social Sciences is an introductory, graduate-level treatment of data analytics for social science. It features applications in the R language, arguably the fastest growing and leading statistical tool for researchers. The book starts with an ethics chapter on the uses and potential abuses of data analytics. Chapters 2 and 3 show how to implement a broad range of statistical procedures in R. Chapters 4 and 5 deal with regression and classification trees and with random forests. Chapter 6 deals with machine learning models and the "caret" package, which makes available to the researcher hundreds of models. Chapter 7 deals with neural network analysis, and Chapter 8 deals with network analysis and visualization of network data. A final chapter treats text analysis, including web scraping, comparative word frequency tables, word clouds, word maps, sentiment analysis, topic analysis, and more. All empirical chapters have two "Quick Start" exercises designed to allow quick immersion in chapter topics, followed by "In Depth" coverage. Data are available for all examples and runnable R code is provided in a "Command Summary". An appendix provides an extended tutorial on R and RStudio. Almost 30 online supplements provide information for the complete book, "books within the book" on a variety of topics, such as agent-based modeling. Rather than focusing on equations, derivations, and proofs, this book emphasizes hands-on obtaining of output for various social science models and how to interpret the output. It is suitable for all advanced level undergraduate and graduate students learning statistical data analysis.
A Strong Practical Focus on Applications and AlgorithmsComputational Statistics Handbook with MATLAB (R), Third Edition covers today's most commonly used techniques in computational statistics while maintaining the same philosophy and writing style of the bestselling previous editions. The text keeps theoretical concepts to a minimum, emphasizing the implementation of the methods. New to the Third EditionThis third edition is updated with the latest version of MATLAB and the corresponding version of the Statistics and Machine Learning Toolbox. It also incorporates new sections on the nearest neighbor classifier, support vector machines, model checking and regularization, partial least squares regression, and multivariate adaptive regression splines. Web ResourceThe authors include algorithmic descriptions of the procedures as well as examples that illustrate the use of algorithms in data analysis. The MATLAB code, examples, and data sets are available online.
* Provides a comprehensive review of methods and applications of Bayesian variable selection. * Divided into four parts: Spike-and-Slab Priors; Continuous Shrinkage Priors; Extensions to various Modeling; Other Approaches to Bayesian Variable Selection. * Covers theoretical and methodological aspects, as well as worked out examples with R code provided in the online supplement. * Includes contributions by experts in the field.
This book reviews nonparametric Bayesian methods and models that have proven useful in the context of data analysis. Rather than providing an encyclopedic review of probability models, the book's structure follows a data analysis perspective. As such, the chapters are organized by traditional data analysis problems. In selecting specific nonparametric models, simpler and more traditional models are favored over specialized ones. The discussed methods are illustrated with a wealth of examples, including applications ranging from stylized examples to case studies from recent literature. The book also includes an extensive discussion of computational methods and details on their implementation. R code for many examples is included in online software pages.
The idea of the Grobner basis first appeared in a 1927 paper by F. S. Macaulay, who succeeded in creating a combinatorial characterization of the Hilbert functions of homogeneous ideals of the polynomial ring. Later, the modern definition of the Grobner basis was independently introduced by Heisuke Hironaka in 1964 and Bruno Buchberger in 1965. However, after the discovery of the notion of the Grobner basis by Hironaka and Buchberger, it was not actively pursued for 20 years. A breakthrough was made in the mid-1980s by David Bayer and Michael Stillman, who created the Macaulay computer algebra system with the help of the Grobner basis. Since then, rapid development on the Grobner basis has been achieved by many researchers, including Bernd Sturmfels. This book serves as a standard bible of the Grobner basis, for which the harmony of theory, application, and computation are indispensable. It provides all the fundamentals for graduate students to learn the ABC s of the Grobner basis, requiring no special knowledgeto understand those basic points. Starting from the introductory performance of the Grobner basis (Chapter 1), a trip around mathematical software follows (Chapter 2). Then comes a deep discussion of how to compute the Grobner basis (Chapter 3). These three chapters may be regarded as the first act of a mathematical play. The second act opens with topics on algebraic statistics (Chapter 4), a fascinating research area where the Grobner basis of a toric ideal is a fundamental tool of the Markov chain Monte Carlo method. Moreover, the Grobner basis of a toric ideal has had a great influence on the study of convex polytopes (Chapter 5). In addition, the Grobner basis of the ring of differential operators gives effective algorithms on holonomic functions (Chapter 6). The third act (Chapter 7) is a collection of concrete examples and problems for Chapters 4, 5 and 6 emphasizing computation by using various software systems."
An introduction to the Central Dogma of molecular biology and information flow in biological systems. A systematic overview of the methods for generating gene expression data. Background knowledge on statistical modeling and machine learning techniques. Detailed methodology of analyzing gene expression data with an example case study. Clustering methods for finding co-expression patterns from microarray, bulkRNA and scRNA data. A large number of practical tools, systems and repositories that are useful for computational biologists to create, analyze and validate biologically relevant gene expression patterns. Suitable for multi-disciplinary researchers and practitioners in computer science and biological sciences.
R is a powerful and free software system for data analysis and graphics, with over 5,000 add-on packages available. This book introduces R using SAS and SPSS terms with which you are already familiar. It demonstrates which of the add-on packages are most like SAS and SPSS and compares them to R's built-in functions. It steps through over 30 programs written in all three packages, comparing and contrasting the packages' differing approaches. The programs and practice datasets are available for download. The glossary defines over 50 R terms using SAS/SPSS jargon and again using R jargon. The table of contents and the index allow you to find equivalent R functions by looking up both SAS statements and SPSS commands. When finished, you will be able to import data, manage and transform it, create publication quality graphics, and perform basic statistical analyses. This new edition has updated programming, an expanded index, and even more statistical methods covered in over 25 new sections.
This book primarily addresses the optimality aspects of covariate designs. A covariate model is a combination of ANOVA and regression models. Optimal estimation of the parameters of the model using a suitable choice of designs is of great importance; as such choices allow experimenters to extract maximum information for the unknown model parameters. The main emphasis of this monograph is to start with an assumed covariate model in combination with some standard ANOVA set-ups such as CRD, RBD, BIBD, GDD, BTIBD, BPEBD, cross-over, multi-factor, split-plot and strip-plot designs, treatment control designs, etc. and discuss the nature and availability of optimal covariate designs. In some situations, optimal estimations of both ANOVA and the regression parameters are provided. Global optimality and D-optimality criteria are mainly used in selecting the design. The standard optimality results of both discrete and continuous set-ups have been adapted, and several novel combinatorial techniques have been applied for the construction of optimum designs using Hadamard matrices, the Kronecker product, Rao-Khatri product, mixed orthogonal arrays to name a few.
The book covers a wide range of topics, yet essential, in Computational Finance (CF), understood as a mix of Finance, Computational Statistics, and Mathematics of Finance. In that regard it is unique in its kind, for it touches upon the basic principles of all three main components of CF, with hands-on examples for programming models in R. Thus, the first chapter gives an introduction to the Principles of Corporate Finance: the markets of stock and options, valuation and economic theory, framed within Computation and Information Theory (e.g. the famous Efficient Market Hypothesis is stated in terms of computational complexity, a new perspective). Chapters 2 and 3 give the necessary tools of Statistics for analyzing financial time series, it also goes in depth into the concepts of correlation, causality and clustering. Chapters 4 and 5 review the most important discrete and continuous models for financial time series. Each model is provided with an example program in R. Chapter 6 covers the essentials of Technical Analysis (TA) and Fundamental Analysis. This chapter is suitable for people outside academics and into the world of financial investments, as a primer in the methods of charting and analysis of value for stocks, as it is done in the financial industry. Moreover, a mathematical foundation to the seemly ad-hoc methods of TA is given, and this is new in a presentation of TA. Chapter 7 reviews the most important heuristics for optimization: simulated annealing, genetic programming, and ant colonies (swarm intelligence) which is material to feed the computer savvy readers. Chapter 8 gives the basic principles of portfolio management, through the mean-variance model, and optimization under different constraints which is a topic of current research in computation, due to its complexity. One important aspect of this chapter is that it teaches how to use the powerful tools for portfolio analysis from the RMetrics R-package. Chapter 9 is a natural continuation of chapter 8 into the new area of research of online portfolio selection. The basic model of the universal portfolio of Cover and approximate methods to computeare alsodescribed."
This book presents selected peer-reviewed contributions from the International Work-Conference on Time Series, ITISE 2017, held in Granada, Spain, September 18-20, 2017. It discusses topics in time series analysis and forecasting, including advanced mathematical methodology, computational intelligence methods for time series, dimensionality reduction and similarity measures, econometric models, energy time series forecasting, forecasting in real problems, online learning in time series as well as high-dimensional and complex/big data time series. The series of ITISE conferences provides a forum for scientists, engineers, educators and students to discuss the latest ideas and implementations in the foundations, theory, models and applications in the field of time series analysis and forecasting. It focuses on interdisciplinary and multidisciplinary research encompassing computer science, mathematics, statistics and econometrics.
In April 2021, the Court of Appeal quashed the convictions of 39 former Subpostmasters and ruled their prosecutions were an affront to the public conscience. They were just a few of the hundreds who had been prosecuted by the Post Office using IT evidence from an unreliable computer system called Horizon. When the Post Office became aware that Horizon didn't work properly, it covered it up. The Great Post Office Scandal is the story of how these innocent people had their lives ruined by a once-loved national institution and how, against overwhelming odds, they fought back to clear their names. Gripping, heart-breaking and enlightening, The Great Post Office Scandal should be read by everyone who wants to understand how this massive miscarriage of justice was allowed to happen.
"MATLAB By Example" guides the reader through each step of writing MATLAB programs. The book assumes no previous programming experience on the part of the reader, and uses multiple examples in clear language to introduce concepts and practical tools. Straightforward and detailed instructions allow beginners to learn and develop their MATLAB skills quickly. The book consists of ten chapters, discussing in detail the
integrated development environment (IDE), scalars, vectors, arrays,
adopting structured programming style using functions and recursive
functions, control flow, debugging, profiling, and structures. A
chapter also describes Symbolic Math Toolbox, teaching readers how
to solve algebraic equations, differentiation, integration,
differential equations, and Laplace and Fourier transforms.
Containing hundreds of examples illustrated using screen shots,
hundreds of exercises, and three projects, this book can be used to
complement coursework or as a self-study book, and can be used as a
textbook in universities, colleges and high schools.
Progressively more and more attention has been paid to how location affects health outcomes. The area of disease mapping focusses on these problems, and the Bayesian paradigm has a major role to play in the understanding of the complex interplay of context and individual predisposition in such studies of disease. Using R for Bayesian Spatial and Spatio-Temporal Health Modeling provides a major resource for those interested in applying Bayesian methodology in small area health data studies. Features: Review of R graphics relevant to spatial health data Overview of Bayesian methods and Bayesian hierarchical modeling as applied to spatial data Bayesian Computation and goodness-of-fit Review of basic Bayesian disease mapping models Spatio-temporal modeling with MCMC and INLA Special topics include multivariate models, survival analysis, missing data, measurement error, variable selection, individual event modeling, and infectious disease modeling Software for fitting models based on BRugs, Nimble, CARBayes and INLA Provides code relevant to fitting all examples throughout the book at a supplementary website The book fills a void in the literature and available software, providing a crucial link for students and professionals alike to engage in the analysis of spatial and spatio-temporal health data from a Bayesian perspective using R. The book emphasizes the use of MCMC via Nimble, BRugs, and CARBAyes, but also includes INLA for comparative purposes. In addition, a wide range of packages useful in the analysis of geo-referenced spatial data are employed and code is provided. It will likely become a key reference for researchers and students from biostatistics, epidemiology, public health, and environmental science.
This book examines current topics and trends in strategic auditing, accounting and finance in digital transformation both from a theoretical and practical perspective. It covers areas such as internal control, corporate governance, enterprise risk management, sustainability and competition. The contributors of this volume emphasize how strategic approaches in this area help companies in achieving targets. The contributions illustrate how by providing good governance, reliable financial reporting, and accountability, businesses can win a competitive advantage. It further discusses how new technological developments like artificial intelligence (AI), cybersystems, network technologies, financial mobility and smart applications, will shape the future of accounting and auditing for firms.
Focused on practical matters: this book will not cover Shiny concepts, but practical tools and methodologies to use for production. Based on experience: this book will be a formalization of several years of experience building Shiny applications. Original content: this book will present new methodology and tooling, not just do a review of what already exists.
Little known to many, R works just as well with JavaScript-this book delves into the various ways both languages can work together. The ultimate aim of this work is to put the reader at ease with inviting JavaScript in their data science workflow. In that respect the book is not teaching one JavaScript but rather we show how little JavaScript can greatly support and enhance R code. Therefore, the focus is on integrating external JavaScript libraries and no prior knowledge of JavaScript is required. Key Features: Easy to pick up. An entry way to learning JavaScript for R. Covers topics not covered anywhere else. Easy to follow along.
Physically Unclonable Functions (PUFs) translate unavoidable variations in certain parameters of materials, waves, or devices into random and unique signals. They have found many applications in the Internet of Things (IoT), authentication systems, FPGA industry, several other areas in communications and related technologies, and many commercial products. Statistical Trend Analysis of Physically Unclonable Functions first presents a review on cryptographic hardware and hardware-assisted cryptography. The review highlights PUF as a mega trend in research on cryptographic hardware design. Afterwards, the authors present a combined survey and research work on PUFs using a systematic approach. As part of the survey aspect, a state-of-the-art analysis is presented as well as a taxonomy on PUFs, a life cycle, and an established ecosystem for the technology. In another part of the survey, the evolutionary history of PUFs is examined, and strategies for further research in this area are suggested. In the research side, this book presents a novel approach for trend analysis that can be applied to any technology or research area. In this method, a text mining tool is used which extracts 1020 keywords from the titles of the sample papers. Then, a classifying tool classifies the keywords into 295 meaningful research topics. The popularity of each topic is then numerically measured and analyzed over the course of time through a statistical analysis on the number of research papers related to the topic as well as the number of their citations. The authors identify the most popular topics in four different domains; over the history of PUFs, during the recent years, in top conferences, and in top journals. The results are used to present an evolution study as well as a trend analysis and develop a roadmap for future research in this area. This method gives an automatic popularity-based statistical trend analysis which eliminates the need for passing personal judgments about the direction of trends, and provides concrete evidence to the future direction of research on PUFs. Another advantage of this method is the possibility of studying a whole lot of existing research works (more than 700 in this book). This book will appeal to researchers in text mining, cryptography, hardware security, and IoT. |
You may like...
Applied Statistics for Environmental…
Abbas F. M. Alkarkhi, Wasin A. A. Alqaraghuli
Paperback
R2,440
Discovery Miles 24 400
Composite Sampling - A Novel Method to…
Ganapati P. Patil, Sharad D. Gore, …
Hardcover
R2,804
Discovery Miles 28 040
Higher Education 4.0 - The Digital…
Kevin Anthony Jones, Sharma Ravishankar
Hardcover
R4,271
Discovery Miles 42 710
Statistical Analysis for…
Arnoldo Frigessi, Peter Buhlmann, …
Hardcover
New Developments in Statistical…
Zhezhen Jin, Mengling Liu, …
Hardcover
R4,614
Discovery Miles 46 140
Model-Free Prediction and Regression - A…
Dimitris N. Politis
Hardcover
R3,041
Discovery Miles 30 410
An Introduction to Clustering with R
Paolo Giordani, Maria Brigida Ferraro, …
Hardcover
R4,002
Discovery Miles 40 020
|