![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer software packages > Other software packages > Mathematical & statistical software
Stata is the most flexible and extensible data analysis package available from a commercial vendor. R is a similarly flexible free and open source package for data analysis, with over 3,000 add-on packages available. This book shows you how to extend the power of Stata through the use of R. It introduces R using Stata terminology with which you are already familiar. It steps through more than 30 programs written in both languages, comparing and contrasting the two packages' different approaches. When finished, you will be able to use R in conjunction with Stata, or separately, to import data, manage and transform it, create publication quality graphics, and perform basic statistical analyses. A glossary defines over 50 R terms using Stata jargon and again using more formal R terminology. The table of contents and index allow you to find equivalent R functions by looking up Stata commands and vice versa. The example programs and practice datasets for both R and Stata are available for download.
This book is a text for a one-semester course for upper-level undergraduates and beginning graduate students in engineering, science, and mathematics. Prerequisites are a first course in the theory of ODEs and a survey course in numerical analysis, in addition to specific programming experience, preferably in MATLAB, and knowledge of elementary matrix theory. Professionals will also find that this useful concise reference contains reviews of technical issues and realistic and detailed examples. The programs for the examples are supplied on the accompanying web site and can serve as templates for solving other problems. Each chapter begins with a discussion of the "facts of life" for the problem, mainly by means of examples. Numerical methods for the problem are then developed, but only those methods most widely used. The treatment of each method is brief and technical issues are minimized, but all the issues important in practice and for understaning the codes are discussed. The last part of each chapter is a tutorial that shows how to solve problems by means of small, but realistic, examples.
Computational inference is based on an approach to statistical methods that uses modern computational power to simulate distributional properties of estimators and test statistics. This book describes computationally intensive statistical methods in a unified presentation, emphasizing techniques, such as the PDF decomposition, that arise in a wide range of methods.
Bayesian statistical methods have become widely used for data analysis and modelling in recent years, and the BUGS software has become the most popular software for Bayesian analysis worldwide. Authored by the team that originally developed this software, The BUGS Book provides a practical introduction to this program and its use. The text presents complete coverage of all the functionalities of BUGS, including prediction, missing data, model criticism, and prior sensitivity. It also features a large number of worked examples and a wide range of applications from various disciplines. The book introduces regression models, techniques for criticism and comparison, and a wide range of modelling issues before going into the vital area of hierarchical models, one of the most common applications of Bayesian methods. It deals with essentials of modelling without getting bogged down in complexity. The book emphasises model criticism, model comparison, sensitivity analysis to alternative priors, and thoughtful choice of prior distributions-all those aspects of the "art" of modelling that are easily overlooked in more theoretical expositions. More pragmatic than ideological, the authors systematically work through the large range of "tricks" that reveal the real power of the BUGS software, for example, dealing with missing data, censoring, grouped data, prediction, ranking, parameter constraints, and so on. Many of the examples are biostatistical, but they do not require domain knowledge and are generalisable to a wide range of other application areas. Full code and data for examples, exercises, and some solutions can be found on the book's website.
Metrics are a hot topic. Executive leadership, boards of directors, management, and customers are all asking for data-based decisions. As a result, many managers, professionals, and change agents are asked to develop metrics, but have no clear idea of how to produce meaningful ones. Wouldn't it be great to have a simple explanation of how to collect, analyze, report, and use measurements to improve your organization? Metrics: How to Improve Key Business Results provides that explanation and the tools you'll need to make your organization more effective. Not only does the book explain the why of metrics, but it walks you through a step-by-step process for creating a report card that provides a clear picture of organizational health and how well you satisfy customer needs. Metrics will help you to measure the right things, the right way - the first time. No wasted effort, no chasing data. The report card provides a simple tool for viewing the health of your organization, from the outside in.You will learn how to measure the key components of the report card and thereby improve real measures of business success, like repeat customers, customer loyalty, and word-of-mouth advertising.This book: * Provides a step-by-step guide for building an organizational effectiveness report card * Takes you from identifying key services and products and using metrics, to determining business strategy * Provides examples of how to identify, collect, analyze, and report metrics that will be immediately useful for improving all aspects of the enterprise, including IT What you'll learn * Understand the difference between data, measures, information, and metrics * Identify root performance questions to ensure you build the right metrics * Develop meaningful and accurate metrics using concrete, easy-to-follow instructions * Avoid the high risks that come with collecting, analyzing, reporting, and using complex data * Formulate practical answers to data-based questions * Select and use the proper tools for creating, implementing, and using metrics * Learn one of the most powerful methods yet invented for improving organizational results Who this book is for Metrics: How to Improve Key Business Results was written for senior managers who need to improve key results.Equally, the book is for the department heads, middle managers, analysts, IT professionals, and change agents responsible for collecting, analyzing, and reporting metrics. Finally, it's for those who have to chase data and find meaningful answers to the interesting questions executives ponder. Table of Contents * Introduction: Who, What, Where, When, Why, and How You Use Metrics * Establishing a Common Language * Where to Begin: Planning a Good Metric * Using Metrics as Indicators * Using the Answer Key * Start with Effectiveness * Triangulation: Essential to Creating Effective Metrics * Expectations: How to View Data in a Meaningful Way * Creating and Interpreting the Metrics Report Card * Final Product: the Metrics Report Card * Employing Advanced Metrics * Creating the Service Catalog * Establishing Standards and Benchmarks * Respecting the Power of Metrics * Avoiding the Research Trap * Embracing Your Organization's Uniqueness * Appendix: Metrics Tools to Use and Useful Resources
Data mining is the art and science of intelligent data analysis. By building knowledge from information, data mining adds considerable value to the ever increasing stores of electronic data that abound today. In performing data mining many decisions need to be made regarding the choice of methodology, the choice of data, the choice of tools, and the choice of algorithms. Throughout this book the reader is introduced to the basic concepts and some of the more popular algorithms of data mining. With a focus on the hands-on end-to-end process for data mining, Williams guides the reader through various capabilities of the easy to use, free, and open source Rattle Data Mining Software built on the sophisticated R Statistical Software. The focus on doing data mining rather than just reading about data mining is refreshing. The book covers data understanding, data preparation, data refinement, model building, model evaluation, and practical deployment. The reader will learn to rapidly deliver a data mining project using software easily installed for free from the Internet. Coupling Rattle with R delivers a very sophisticated data mining environment with all the power, and more, of the many commercial offerings.
Advanced R helps you understand how R works at a fundamental level. It is designed for R programmers who want to deepen their understanding of the language, and programmers experienced in other languages who want to understand what makes R different and special. This book will teach you the foundations of R; three fundamental programming paradigms (functional, object-oriented, and metaprogramming); and powerful techniques for debugging and optimising your code. By reading this book, you will learn: The difference between an object and its name, and why the distinction is important The important vector data structures, how they fit together, and how you can pull them apart using subsetting The fine details of functions and environments The condition system, which powers messages, warnings, and errors The powerful functional programming paradigm, which can replace many for loops The three most important OO systems: S3, S4, and R6 The tidy eval toolkit for metaprogramming, which allows you to manipulate code and control evaluation Effective debugging techniques that you can deploy, regardless of how your code is run How to find and remove performance bottlenecks The second edition is a comprehensive update: New foundational chapters: "Names and values," "Control flow," and "Conditions" comprehensive coverage of object oriented programming with chapters on S3, S4, R6, and how to choose between them Much deeper coverage of metaprogramming, including the new tidy evaluation framework use of new package like rlang (http://rlang.r-lib.org), which provides a clean interface to low-level operations, and purr (http://purrr.tidyverse.org/) for functional programming Use of color in code chunks and figures Hadley Wickham is Chief Scientist at RStudio, an Adjunct Professor at Stanford University and the University of Auckland, and a member of the R Foundation. He is the lead developer of the tidyverse, a collection of R packages, including ggplot2 and dplyr, designed to support data science. He is also the author of R for Data Science (with Garrett Grolemund), R Packages, and ggplot2: Elegant Graphics for Data Analysis.
There has been dramatic growth in the development and application of Bayesian inference in statistics. Berger (2000) documents the increase in Bayesian activity by the number of published research articles, the number of books, andtheextensivenumberofapplicationsofBayesianarticlesinapplied disciplines such as science and engineering. One reason for the dramatic growth in Bayesian modeling is the availab- ity of computational algorithms to compute the range of integrals that are necessary in a Bayesian posterior analysis. Due to the speed of modern c- puters, it is now possible to use the Bayesian paradigm to 't very complex models that cannot be 't by alternative frequentist methods. To 't Bayesian models, one needs a statistical computing environment. This environment should be such that one can: write short scripts to de?ne a Bayesian model use or write functions to summarize a posterior distribution use functions to simulate from the posterior distribution construct graphs to illustrate the posterior inference An environment that meets these requirements is the R system. R provides a wide range of functions for data manipulation, calculation, and graphical d- plays. Moreover, it includes a well-developed, simple programming language that users can extend by adding new functions. Many such extensions of the language in the form of packages are easily downloadable from the Comp- hensive R Archive Network (CRAN)
BRIDGES THE GAP BETWEEN SAS AND R, ALLOWING USERS TRAINED IN ONE LANGUAGE TO EASILY LEARN THE OTHER SAS and R are widely-used, very different software environments. Prized for its statistical and graphical tools, R is an open-source programming language that is popular with statisticians and data miners who develop statistical software and analyze data. SAS (Statistical Analysis System) is the leading corporate software in analytics thanks to its faster data handling and smaller learning curve. SAS for R Users enables entry-level data scientists to take advantage of the best aspects of both tools by providing a cross-functional framework for users who already know R but may need to work with SAS. Those with knowledge of both R and SAS are of far greater value to employers, particularly in corporate settings. Using a clear, step-by-step approach, this book presents an analytics workflow that mirrors that of the everyday data scientist. This up-to-date guide is compatible with the latest R packages as well as SAS University Edition. Useful for anyone seeking employment in data science, this book: Instructs both practitioners and students fluent in one language seeking to learn the other Provides command-by-command translations of R to SAS and SAS to R Offers examples and applications in both R and SAS Presents step-by-step guidance on workflows, color illustrations, sample code, chapter quizzes, and more Includes sections on advanced methods and applications Designed for professionals, researchers, and students, SAS for R Users is a valuable resource for those with some knowledge of coding and basic statistics who wish to enter the realm of data science and business analytics.
Go from total MATLAB newbie to plotting graphs and solving equations in a flash! MATLAB is one of the most powerful and commonly used tools in the STEM field. But did you know it doesn't take an advanced degree or a ton of computer experience to learn it? MATLAB For Dummies is the roadmap you've been looking for to simplify and explain this feature-filled tool. This handy reference walks you through every step of the way as you learn the MATLAB language and environment inside-and-out. Starting with straightforward basics before moving on to more advanced material like Live Functions and Live Scripts, this easy-to-read guide shows you how to make your way around MATLAB with screenshots and newly updated procedures. It includes: A comprehensive introduction to installing MATLAB, using its interface, and creating and saving your first file Fully updated to include the 2020 and 2021 updates to MATLAB, with all-new screenshots and up-to-date procedures Enhanced debugging procedures and use of the Symbolic Math Toolbox Brand new instruction on working with Live Scripts and Live Functions, designing classes, creating apps, and building projects Intuitive walkthroughs for MATLAB's advanced features, including importing and exporting data and publishing your work Perfect for STEM students and new professionals ready to master one of the most powerful tools in the fields of engineering, mathematics, and computing, MATLAB For Dummies is the simplest way to go from complete newbie to power user faster than you would have thought possible.
Can artificial intelligence learn mathematics? The question is at the heart of this original monograph bringing together theoretical physics, modern geometry, and data science. The study of Calabi-Yau manifolds lies at an exciting intersection between physics and mathematics. Recently, there has been much activity in applying machine learning to solve otherwise intractable problems, to conjecture new formulae, or to understand the underlying structure of mathematics. In this book, insights from string and quantum field theory are combined with powerful techniques from complex and algebraic geometry, then translated into algorithms with the ultimate aim of deriving new information about Calabi-Yau manifolds. While the motivation comes from mathematical physics, the techniques are purely mathematical and the theme is that of explicit calculations. The reader is guided through the theory and provided with explicit computer code in standard software such as SageMath, Python and Mathematica to gain hands-on experience in applications of artificial intelligence to geometry. Driven by data and written in an informal style, The Calabi-Yau Landscape makes cutting-edge topics in mathematical physics, geometry and machine learning readily accessible to graduate students and beyond. The overriding ambition is to introduce some modern mathematics to the physicist, some modern physics to the mathematician, and machine learning to both.
R is a language and environment for data analysis and graphics. It may be considered an implementation of S, an award-winning language initially - veloped at Bell Laboratories since the late 1970s. The R project was initiated by Robert Gentleman and Ross Ihaka at the University of Auckland, New Zealand, in the early 1990s, and has been developed by an international team since mid-1997. Historically, econometricians have favored other computing environments, some of which have fallen by the wayside, and also a variety of packages with canned routines. We believe that R has great potential in econometrics, both for research and for teaching. There are at least three reasons for this: (1) R is mostly platform independent and runs on Microsoft Windows, the Mac family of operating systems, and various ?avors of Unix/Linux, and also on some more exotic platforms. (2) R is free software that can be downloaded and installed at no cost from a family of mirror sites around the globe, the Comprehensive R Archive Network (CRAN); hence students can easily install it on their own machines. (3) R is open-source software, so that the full source code is available and can be inspected to understand what it really does, learn from it, and modify and extend it. We also like to think that platform independence and the open-source philosophy make R an ideal environment for reproducible econometric research.
"Modeling with Data" fully explains how to execute computationally intensive analyses on very large data sets, showing readers how to determine the best methods for solving a variety of different problems, how to create and debug statistical models, and how to run an analysis and evaluate the results. Ben Klemens introduces a set of open and unlimited tools, and uses them to demonstrate data management, analysis, and simulation techniques essential for dealing with large data sets and computationally intensive procedures. He then demonstrates how to easily apply these tools to the many threads of statistical technique, including classical, Bayesian, maximum likelihood, and Monte Carlo methods. Klemens's accessible survey describes these models in a unified and nontraditional manner, providing alternative ways of looking at statistical concepts that often befuddle students. The book includes nearly one hundred sample programs of all kinds. Links to these programs will be available on this page at a later date. "Modeling with Data" will interest anyone looking for a comprehensive guide to these powerful statistical tools, including researchers and graduate students in the social sciences, biology, engineering, economics, and applied mathematics.
R is rapidly becoming the standard software for statistical analyses, graphical presentation of data, and programming in the natural, physical, social, and engineering sciences. Getting Started with R is now the go-to introductory guide for biologists wanting to learn how to use R in their research. It teaches readers how to import, explore, graph, and analyse data, while keeping them focused on their ultimate goals: clearly communicating their data in oral presentations, posters, papers, and reports. It provides a consistent workflow for using R that is simple, efficient, reliable, and reproducible. This second edition has been updated and expanded while retaining the concise and engaging nature of its predecessor, offering an accessible and fun introduction to the packages dplyr and ggplot2 for data manipulation and graphing. It expands the set of basic statistics considered in the first edition to include new examples of a simple regression, a one-way and a two-way ANOVA. Finally, it introduces a new chapter on the generalised linear model. Getting Started with R is suitable for undergraduates, graduate students, professional researchers, and practitioners in the biological sciences.
"Mathematica Cookbook" helps you master the application's core principles by walking you through real-world problems. Ideal for browsing, this book includes recipes for working with numerics, data structures, algebraic equations, calculus, and statistics. You'll also venture into exotic territory with recipes for data visualization using 2D and 3D graphic tools, image processing, and music. Although Mathematica 7 is a highly advanced computational platform, the recipes in this book make it accessible to everyone -- whether you're working on high school algebra, simple graphs, PhD-level computation, financial analysis, or advanced engineering models.Learn how to use Mathematica at a higher level with functional programming and pattern matchingDelve into the rich library of functions for string and structured text manipulationLearn how to apply the tools to physics and engineering problemsDraw on Mathematica's access to physics, chemistry, and biology dataGet techniques for solving equations in computational financeLearn how to use Mathematica for sophisticated image processingProcess music and audio as musical notes, analog waveforms, or digital sound samples
New statistical tools are changing the ways in which scientists analyze and interpret data and models. Many of these are emerging as a result of the wide availability of inexpensive, high speed computational power. In particular, hierarchical Bayes and Markov Chain Monte Carlo methods for anyalsis provide consistent framework for inference and prediction where information is heterogeneous and uncertain, processes are complex, and responses depend on scale. Nowhere are these methods more promising than in the environmental sciences. Models have developed rapidly, and there is now a requirement for a clear exposition of the methodology through to application for a range of environmental challenges.
This volume contains selected papers presented at two joint German-Japanese symposia on data analysis and related elds. The articles substantially extend and further develop material presented at the two symposia organized on the basis of longstanding and close relationships which have been cultivated in the last couple of decades between the two classi cation societies: the German Class- cation Society (Gesellschaft fu ]r Klassi kation e. V.) and the Japanese Classi cation Society. These symposia have been very helpful in exchanging ideas, views, and knowledge between the two societies and have served as a spring board for more extensive and closer co-operation between the societies as well as among their individual members. The scienti c program of the rst Joint Japanese-German Symposium (Tokyo 2005)included23presentations;forthesecondJointGerman-JapaneseSymposium (Berlin 2006) 27 presentations were scheduled. This volume presents 21 peer refereed papers, which are grouped into three parts: 1. Part 1 Clustering and Visualization (eight papers) 2. Part 2 Methods in Fields (nine papers) 3. Part 3 Applications in Clustering and Visualization (four papers) The concept of having a joint symposium of the two classi cation societies came from the talks with Hans-Hermann and Wolfgang when Akinori attended the 28th Annual Conference of the German Classi cation Society held in Dortmund in March 2004."
International Association for Statistical Computing The International Association for Statistical Computing (IASC) is a Section of the International Statistical Institute. The objectives of the Association are to foster world-wide interest in e?ective statistical computing and to - change technical knowledge through international contacts and meetings - tween statisticians, computing professionals, organizations, institutions, g- ernments and the general public. The IASC organises its own Conferences, IASC World Conferences, and COMPSTAT in Europe. The 17th Conference of ERS-IASC, the biennial meeting of European - gional Section of the IASC was held in Rome August 28 - September 1, 2006. This conference took place in Rome exactly 20 years after the 7th COMP- STAT symposium which was held in Rome, in 1986. Previous COMPSTAT conferences were held in: Vienna (Austria, 1974); West-Berlin (Germany, 1976); Leiden (The Netherlands, 1978); Edimbourgh (UK, 1980); Toulouse (France, 1982); Prague (Czechoslovakia, 1984); Rome (Italy, 1986); Copenhagen (Denmark, 1988); Dubrovnik (Yugoslavia, 1990); Neuch atel (Switzerland, 1992); Vienna (Austria,1994); Barcelona (Spain, 1996);Bristol(UK,1998);Utrecht(TheNetherlands,2000);Berlin(Germany, 2002); Prague (Czech Republic, 2004).
The updated guide to the newest graphing calculator from Texas Instruments The TI-Nspire graphing calculator is popular among high school and college students as a valuable tool for calculus, AP calculus, and college-level algebra courses. Its use is allowed on the major college entrance exams. This book is a nuts-and-bolts guide to working with the TI-Nspire, providing everything you need to get up and running and helping you get the most out of this high-powered math tool.Texas Instruments' TI-Nspire graphing calculator is perfect for high school and college students in advanced algebra and calculus classes as well as students taking the SAT, PSAT, and ACT examsThis fully updated guide covers all enhancements to the TI-Nspire, including the touchpad and the updated software that can be purchased along with the deviceShows how to get maximum value from this versatile math tool With updated screenshots and examples, "TI-Nspire For Dummies" provides practical, hands-on instruction to help students make the most of this revolutionary graphing calculator.
An elementary first course for students in mathematics and engineering Practical in approach: examples of code are provided for students to debug, and tasks - with full solutions - are provided at the end of each chapter Includes a glossary of useful terms, with each term supported by an example of the syntaxes commonly encountered
|
You may like...
Essential Java for Scientists and…
Brian Hahn, Katherine Malan
Paperback
R1,266
Discovery Miles 12 660
An Introduction to Creating Standardized…
Todd Case, Yuting Tian
Hardcover
R1,501
Discovery Miles 15 010
Mathematical Modeling for Smart…
Debabrata Samanta, Debabrata Singh
Hardcover
R11,427
Discovery Miles 114 270
Jump into JMP Scripting, Second Edition…
Wendy Murphrey, Rosemary Lucas
Hardcover
R1,530
Discovery Miles 15 300
|