![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer software packages > Other software packages > Mathematical & statistical software
An overview of the theory and application of linear and nonlinear mixed-effects models in the analysis of grouped data, such as longitudinal data, repeated measures, and multilevel data. The authors present a unified model-building strategy for both models and apply this to the analysis of over 20 real datasets from a wide variety of areas, including pharmacokinetics, agriculture, and manufacturing. Much emphasis is placed on the use of graphical displays at the various phases of the model-building process, starting with exploratory plots of the data and concluding with diagnostic plots to assess the adequacy of a fitted model. The NLME library for analyzing mixed-effects models in S and S-PLUS, developed by the authors, provides the underlying software for implementing the methods presented. This balanced mix of real data examples, modeling software, and theory makes the book a useful reference for practitioners who use, or intend to use, mixed-effects models in their data analyses. It can also be used as a text for a one-semester graduate-level applied course.
Discover the story of your data using the essential elements of associations and correlations Key Features Get a comprehensive introduction to associations and correlations Explore multivariate analysis, understand its limitations, and discover the assumptions on which it's based Gain insights into the various ways of preparing your data for analysis and visualization Book DescriptionAssociations and correlations are ways of describing how a pair of variables change together as a result of their connection. By knowing the various available techniques, you can easily and accurately discover and visualize the relationships in your data. This book begins by showing you how to classify your data into the four distinct types that you are likely to have in your dataset. Then, with easy-to-understand examples, you'll learn when to use the various univariate and multivariate statistical tests. You'll also discover what to do when your univariate and multivariate results do not match. As the book progresses, it describes why univariate and multivariate techniques should be used as a tag team, and also introduces you to the techniques of visualizing the story of your data. By the end of the book, you'll know exactly how to select the most appropriate univariate and multivariate tests, and be able to use a single strategic framework to discover the true story of your data. What you will learn Identify a dataset that's fit for analysis using its basic features Understand the importance of associations and correlations Use multivariate and univariate statistical tests to confirm relationships Classify data as qualitative or quantitative and then into the four subtypes Build a visual representation of all the relationships in the dataset Automate associations and correlations with CorrelViz Who this book is forThis is a book for beginners - if you're a novice data analyst or data scientist, then this is a great place to start. Experienced data analysts might also find value in this title, as it will recap the basics and strengthen your understanding of key concepts. This book focuses on introducing the essential elements of association and correlation analysis.
Hurricanes are nature's most destructive storms and they are becoming more powerful as the globe warms. Hurricane Climatology explains how to analyze and model hurricane data to better understand and predict present and future hurricane activity. It uses the open-source and now widely used R software for statistical computing to create a tutorial-style manual for independent study, review, and reference. The text is written around the code that when copied will reproduce the graphs, tables, and maps. The approach is different from other books that use R. It focuses on a single topic and explains how to make use of R to better understand the topic. The book is organized into two parts, the first of which provides material on software, statistics, and data. The second part presents methods and models used in hurricane climate research.
This is the first book to show the capabilities of Microsoft Excel to teach biological and life sciences statistics effectively. It is a step-by-step exercise-driven guide for students and practitioners who need to master Excel to solve practical science problems. If understanding statistics isn t your strongest suit, you are not especially mathematically-inclined, or if you are wary of computers, this is the right book for you. Excel, a widely available computer program for students and managers, is also an effective teaching and learning tool for quantitative analyses in science courses. Its powerful computational ability and graphical functions make learning statistics much easier than in years past. However, "Excel 2010 for Biological and Life Sciences Statistics: A Guide to Solving Practical Problems" is the first book to capitalize on these improvements by teaching students and managers how to apply Excel to statistical techniques necessary in their courses and work. Each chapter explains statistical formulas and directs the reader to use Excel commands to solve specific, easy-to-understand science problems. Practice problems are provided at the end of each chapter with their solutions in an appendix. Separately, there is a full Practice Test (with answers in an Appendix) that allows readers to test what they have learned. "
Six Sigma has arisen in the last two decades as a breakthrough Quality Management Methodology. With Six Sigma, we are solving problems and improving processes using as a basis one of the most powerful tools of human development: the scientific method. For the analysis of data, Six Sigma requires the use of statistical software, being R an Open Source option that fulfills this requirement. R is a software system that includes a programming language widely used in academic and research departments. Nowadays, it is becoming a real alternative within corporate environments. The aim of this book is to show how R can be used as the software tool in the development of Six Sigma projects. The book includes a gentle introduction to Six Sigma and a variety of examples showing how to use R within real situations. It has been conceived as a self contained piece. Therefore, it is addressed not only to Six Sigma practitioners, but also to professionals trying to initiate themselves in this management methodology. The book may be used as a text book as well.
The richly illustrated Interactive Web-Based Data Visualization with R, plotly, and shiny focuses on the process of programming interactive web graphics for multidimensional data analysis. It is written for the data analyst who wants to leverage the capabilities of interactive web graphics without having to learn web programming. Through many R code examples, you will learn how to tap the extensive functionality of these tools to enhance the presentation and exploration of data. By mastering these concepts and tools, you will impress your colleagues with your ability to quickly generate more informative, engaging, and reproducible interactive graphics using free and open source software that you can share over email, export to pdf, and more. Key Features: Convert static ggplot2 graphics to an interactive web-based form Link, animate, and arrange multiple plots in standalone HTML from R Embed, modify, and respond to plotly graphics in a shiny app Learn best practices for visualizing continuous, discrete, and multivariate data Learn numerous ways to visualize geo-spatial data This book makes heavy use of plotly for graphical rendering, but you will also learn about other R packages that support different phases of a data science workflow, such as tidyr, dplyr, and tidyverse. Along the way, you will gain insight into best practices for visualization of high-dimensional data, statistical graphics, and graphical perception. The printed book is complemented by an interactive website where readers can view movies demonstrating the examples and interact with graphics.
This book evolved from lectures, courses and workshops on missing data and small-area estimation that I presented during my tenure as the ?rst C- pion Fellow (2000-2002). For the Fellowship I proposed these two topics as areas in which the academic statistics could contribute to the development of government statistics, in exchange for access to the operational details and background that would inform the direction and sharpen the focus of a- demic research. After a few years of involvement, I have come to realise that the separation of 'academic' and 'industrial' statistics is not well suited to either party, and their integration is the key to progress in both branches. Most of the work on this monograph was done while I was a visiting l- turer at Massey University, Palmerston North, New Zealand. The hospitality and stimulating academic environment of their Institute of Information S- ence and Technology is gratefully acknowledged. I could not name all those who commented on my lecture notes and on the presentations themselves; apart from them, I want to thank the organisers and silent attendees of all the events, and, with a modicum of reluctance, the 'grey ?gures' who kept inquiring whether I was any nearer the completion of whatever stage I had been foolish enough to attach a date.
This book presents a study of the COVID-19 pandemic using computational social scientific analysis that draws from, and employs, statistics and simulations. Combining approaches in crisis management, risk assessment and mathematical modelling, the work also draws from the philosophy of sacrifice and futurology. It makes an original contribution to the important issue of the stability of society by highlighting two significant factors: the COVID-19 crisis as a catalyst for change and the rise of AI and Big Data in managing society. It also emphasizes the nature and importance of sacrifices and the role of politics in the distribution of sacrifices. The book considers the treatment of AI and Big Data and their use to both "good" and "bad" ends, exposing the inevitability of these tools being used. Relevant to both policymakers and social scientists interested in the influence of AI and Big Data on the structure of society, the book re-evaluates the ways we think of lifestyles, economic systems and the balance of power in tandem with digital transformation.
The first part of this title contained all statistical tests that are relevant for starters on SPSS, and included standard parametric and non-parametric tests for continuous and binary variables, regression methods, trend tests, and reliability and validity assessments of diagnostic tests. The current part 2 of this title reviews multistep methods, multivariate models, assessments of missing data, performance of diagnostic tests, meta-regression, Poisson regression, confounding and interaction, and survival analyses using log tests and segmented time-dependent Cox regression. Methods for assessing non linear models, data seasonality, distribution free methods, including Monte Carlo methods and artificial intelligence, and robust tests are also covered. Each method of testing is explained using a data example from clinical practice, including every step in SPSS, and a text with interpretations of the results and hints convenient for data reporting. In order to facilitate the use of this cookbook the data files of the examples is made available by the editor through extras.springer.com. Both part 1 and 2 of this title contain a minima amount of text and maximal technical details, but we believe that this property will not refrain students from mastering the SPSS software systematics, and that, instead, it will be a help to that aim. Yet, we recommend that it will used together with the textbook "Statistics Applied to Clinical Trials" (5th edition, Springer, Dordrecht 2012) and the e-books "Statistics on a Pocket Calculator Part 1 and 2 (Springer, Dordrecht, 2011 and 2012) from the same authors.
Stata is the most flexible and extensible data analysis package available from a commercial vendor. R is a similarly flexible free and open source package for data analysis, with over 3,000 add-on packages available. This book shows you how to extend the power of Stata through the use of R. It introduces R using Stata terminology with which you are already familiar. It steps through more than 30 programs written in both languages, comparing and contrasting the two packages' different approaches. When finished, you will be able to use R in conjunction with Stata, or separately, to import data, manage and transform it, create publication quality graphics, and perform basic statistical analyses. A glossary defines over 50 R terms using Stata jargon and again using more formal R terminology. The table of contents and index allow you to find equivalent R functions by looking up Stata commands and vice versa. The example programs and practice datasets for both R and Stata are available for download.
This book is a text for a one-semester course for upper-level undergraduates and beginning graduate students in engineering, science, and mathematics. Prerequisites are a first course in the theory of ODEs and a survey course in numerical analysis, in addition to specific programming experience, preferably in MATLAB, and knowledge of elementary matrix theory. Professionals will also find that this useful concise reference contains reviews of technical issues and realistic and detailed examples. The programs for the examples are supplied on the accompanying web site and can serve as templates for solving other problems. Each chapter begins with a discussion of the "facts of life" for the problem, mainly by means of examples. Numerical methods for the problem are then developed, but only those methods most widely used. The treatment of each method is brief and technical issues are minimized, but all the issues important in practice and for understaning the codes are discussed. The last part of each chapter is a tutorial that shows how to solve problems by means of small, but realistic, examples.
Computational inference is based on an approach to statistical methods that uses modern computational power to simulate distributional properties of estimators and test statistics. This book describes computationally intensive statistical methods in a unified presentation, emphasizing techniques, such as the PDF decomposition, that arise in a wide range of methods.
Bayesian statistical methods have become widely used for data analysis and modelling in recent years, and the BUGS software has become the most popular software for Bayesian analysis worldwide. Authored by the team that originally developed this software, The BUGS Book provides a practical introduction to this program and its use. The text presents complete coverage of all the functionalities of BUGS, including prediction, missing data, model criticism, and prior sensitivity. It also features a large number of worked examples and a wide range of applications from various disciplines. The book introduces regression models, techniques for criticism and comparison, and a wide range of modelling issues before going into the vital area of hierarchical models, one of the most common applications of Bayesian methods. It deals with essentials of modelling without getting bogged down in complexity. The book emphasises model criticism, model comparison, sensitivity analysis to alternative priors, and thoughtful choice of prior distributions-all those aspects of the "art" of modelling that are easily overlooked in more theoretical expositions. More pragmatic than ideological, the authors systematically work through the large range of "tricks" that reveal the real power of the BUGS software, for example, dealing with missing data, censoring, grouped data, prediction, ranking, parameter constraints, and so on. Many of the examples are biostatistical, but they do not require domain knowledge and are generalisable to a wide range of other application areas. Full code and data for examples, exercises, and some solutions can be found on the book's website.
Metrics are a hot topic. Executive leadership, boards of directors, management, and customers are all asking for data-based decisions. As a result, many managers, professionals, and change agents are asked to develop metrics, but have no clear idea of how to produce meaningful ones. Wouldn't it be great to have a simple explanation of how to collect, analyze, report, and use measurements to improve your organization? Metrics: How to Improve Key Business Results provides that explanation and the tools you'll need to make your organization more effective. Not only does the book explain the why of metrics, but it walks you through a step-by-step process for creating a report card that provides a clear picture of organizational health and how well you satisfy customer needs. Metrics will help you to measure the right things, the right way - the first time. No wasted effort, no chasing data. The report card provides a simple tool for viewing the health of your organization, from the outside in.You will learn how to measure the key components of the report card and thereby improve real measures of business success, like repeat customers, customer loyalty, and word-of-mouth advertising.This book: * Provides a step-by-step guide for building an organizational effectiveness report card * Takes you from identifying key services and products and using metrics, to determining business strategy * Provides examples of how to identify, collect, analyze, and report metrics that will be immediately useful for improving all aspects of the enterprise, including IT What you'll learn * Understand the difference between data, measures, information, and metrics * Identify root performance questions to ensure you build the right metrics * Develop meaningful and accurate metrics using concrete, easy-to-follow instructions * Avoid the high risks that come with collecting, analyzing, reporting, and using complex data * Formulate practical answers to data-based questions * Select and use the proper tools for creating, implementing, and using metrics * Learn one of the most powerful methods yet invented for improving organizational results Who this book is for Metrics: How to Improve Key Business Results was written for senior managers who need to improve key results.Equally, the book is for the department heads, middle managers, analysts, IT professionals, and change agents responsible for collecting, analyzing, and reporting metrics. Finally, it's for those who have to chase data and find meaningful answers to the interesting questions executives ponder. Table of Contents * Introduction: Who, What, Where, When, Why, and How You Use Metrics * Establishing a Common Language * Where to Begin: Planning a Good Metric * Using Metrics as Indicators * Using the Answer Key * Start with Effectiveness * Triangulation: Essential to Creating Effective Metrics * Expectations: How to View Data in a Meaningful Way * Creating and Interpreting the Metrics Report Card * Final Product: the Metrics Report Card * Employing Advanced Metrics * Creating the Service Catalog * Establishing Standards and Benchmarks * Respecting the Power of Metrics * Avoiding the Research Trap * Embracing Your Organization's Uniqueness * Appendix: Metrics Tools to Use and Useful Resources
Data mining is the art and science of intelligent data analysis. By building knowledge from information, data mining adds considerable value to the ever increasing stores of electronic data that abound today. In performing data mining many decisions need to be made regarding the choice of methodology, the choice of data, the choice of tools, and the choice of algorithms. Throughout this book the reader is introduced to the basic concepts and some of the more popular algorithms of data mining. With a focus on the hands-on end-to-end process for data mining, Williams guides the reader through various capabilities of the easy to use, free, and open source Rattle Data Mining Software built on the sophisticated R Statistical Software. The focus on doing data mining rather than just reading about data mining is refreshing. The book covers data understanding, data preparation, data refinement, model building, model evaluation, and practical deployment. The reader will learn to rapidly deliver a data mining project using software easily installed for free from the Internet. Coupling Rattle with R delivers a very sophisticated data mining environment with all the power, and more, of the many commercial offerings.
Advanced R helps you understand how R works at a fundamental level. It is designed for R programmers who want to deepen their understanding of the language, and programmers experienced in other languages who want to understand what makes R different and special. This book will teach you the foundations of R; three fundamental programming paradigms (functional, object-oriented, and metaprogramming); and powerful techniques for debugging and optimising your code. By reading this book, you will learn: The difference between an object and its name, and why the distinction is important The important vector data structures, how they fit together, and how you can pull them apart using subsetting The fine details of functions and environments The condition system, which powers messages, warnings, and errors The powerful functional programming paradigm, which can replace many for loops The three most important OO systems: S3, S4, and R6 The tidy eval toolkit for metaprogramming, which allows you to manipulate code and control evaluation Effective debugging techniques that you can deploy, regardless of how your code is run How to find and remove performance bottlenecks The second edition is a comprehensive update: New foundational chapters: "Names and values," "Control flow," and "Conditions" comprehensive coverage of object oriented programming with chapters on S3, S4, R6, and how to choose between them Much deeper coverage of metaprogramming, including the new tidy evaluation framework use of new package like rlang (http://rlang.r-lib.org), which provides a clean interface to low-level operations, and purr (http://purrr.tidyverse.org/) for functional programming Use of color in code chunks and figures Hadley Wickham is Chief Scientist at RStudio, an Adjunct Professor at Stanford University and the University of Auckland, and a member of the R Foundation. He is the lead developer of the tidyverse, a collection of R packages, including ggplot2 and dplyr, designed to support data science. He is also the author of R for Data Science (with Garrett Grolemund), R Packages, and ggplot2: Elegant Graphics for Data Analysis.
There has been dramatic growth in the development and application of Bayesian inference in statistics. Berger (2000) documents the increase in Bayesian activity by the number of published research articles, the number of books, andtheextensivenumberofapplicationsofBayesianarticlesinapplied disciplines such as science and engineering. One reason for the dramatic growth in Bayesian modeling is the availab- ity of computational algorithms to compute the range of integrals that are necessary in a Bayesian posterior analysis. Due to the speed of modern c- puters, it is now possible to use the Bayesian paradigm to 't very complex models that cannot be 't by alternative frequentist methods. To 't Bayesian models, one needs a statistical computing environment. This environment should be such that one can: write short scripts to de?ne a Bayesian model use or write functions to summarize a posterior distribution use functions to simulate from the posterior distribution construct graphs to illustrate the posterior inference An environment that meets these requirements is the R system. R provides a wide range of functions for data manipulation, calculation, and graphical d- plays. Moreover, it includes a well-developed, simple programming language that users can extend by adding new functions. Many such extensions of the language in the form of packages are easily downloadable from the Comp- hensive R Archive Network (CRAN)
BRIDGES THE GAP BETWEEN SAS AND R, ALLOWING USERS TRAINED IN ONE LANGUAGE TO EASILY LEARN THE OTHER SAS and R are widely-used, very different software environments. Prized for its statistical and graphical tools, R is an open-source programming language that is popular with statisticians and data miners who develop statistical software and analyze data. SAS (Statistical Analysis System) is the leading corporate software in analytics thanks to its faster data handling and smaller learning curve. SAS for R Users enables entry-level data scientists to take advantage of the best aspects of both tools by providing a cross-functional framework for users who already know R but may need to work with SAS. Those with knowledge of both R and SAS are of far greater value to employers, particularly in corporate settings. Using a clear, step-by-step approach, this book presents an analytics workflow that mirrors that of the everyday data scientist. This up-to-date guide is compatible with the latest R packages as well as SAS University Edition. Useful for anyone seeking employment in data science, this book: Instructs both practitioners and students fluent in one language seeking to learn the other Provides command-by-command translations of R to SAS and SAS to R Offers examples and applications in both R and SAS Presents step-by-step guidance on workflows, color illustrations, sample code, chapter quizzes, and more Includes sections on advanced methods and applications Designed for professionals, researchers, and students, SAS for R Users is a valuable resource for those with some knowledge of coding and basic statistics who wish to enter the realm of data science and business analytics.
Computational techniques based on simulation have now become an essential part of the statistician's toolbox. It is thus crucial to provide statisticians with a practical understanding of those methods, and there is no better way to develop intuition and skills for simulation than to use simulation to solve statistical problems. Introducing Monte Carlo Methods with R covers the main tools used in statistical simulation from a programmer's point of view, explaining the R implementation of each simulation technique and providing the output for better understanding and comparison. While this book constitutes a comprehensive treatment of simulation methods, the theoretical justification of those methods has been considerably reduced, compared with Robert and Casella (2004). Similarly, the more exploratory and less stable solutions are not covered here. This book does not require a preliminary exposure to the R programming language or to Monte Carlo methods, nor an advanced mathematical background. While many examples are set within a Bayesian framework, advanced expertise in Bayesian statistics is not required. The book covers basic random generation algorithms, Monte Carlo techniques for integration and optimization, convergence diagnoses, Markov chain Monte Carlo methods, including Metropolis {Hastings and Gibbs algorithms, and adaptive algorithms. All chapters include exercises and all R programs are available as an R package called mcsm. The book appeals to anyone with a practical interest in simulation methods but no previous exposure. It is meant to be useful for students and practitioners in areas such as statistics, signal processing, communications engineering, control theory, econometrics, finance and more. The programming parts are introduced progressively to be accessible to any reader.
Can artificial intelligence learn mathematics? The question is at the heart of this original monograph bringing together theoretical physics, modern geometry, and data science. The study of Calabi-Yau manifolds lies at an exciting intersection between physics and mathematics. Recently, there has been much activity in applying machine learning to solve otherwise intractable problems, to conjecture new formulae, or to understand the underlying structure of mathematics. In this book, insights from string and quantum field theory are combined with powerful techniques from complex and algebraic geometry, then translated into algorithms with the ultimate aim of deriving new information about Calabi-Yau manifolds. While the motivation comes from mathematical physics, the techniques are purely mathematical and the theme is that of explicit calculations. The reader is guided through the theory and provided with explicit computer code in standard software such as SageMath, Python and Mathematica to gain hands-on experience in applications of artificial intelligence to geometry. Driven by data and written in an informal style, The Calabi-Yau Landscape makes cutting-edge topics in mathematical physics, geometry and machine learning readily accessible to graduate students and beyond. The overriding ambition is to introduce some modern mathematics to the physicist, some modern physics to the mathematician, and machine learning to both. |
You may like...
Essential Java for Scientists and…
Brian Hahn, Katherine Malan
Paperback
R1,266
Discovery Miles 12 660
An Introduction to Creating Standardized…
Todd Case, Yuting Tian
Hardcover
R1,501
Discovery Miles 15 010
Mathematical Modeling for Smart…
Debabrata Samanta, Debabrata Singh
Hardcover
R11,427
Discovery Miles 114 270
Jump into JMP Scripting, Second Edition…
Wendy Murphrey, Rosemary Lucas
Hardcover
R1,530
Discovery Miles 15 300
|