![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer software packages > Other software packages > Mathematical & statistical software
Complex Analysis with Mathematica offers a new way of learning and teaching a subject that lies at the heart of many areas of pure and applied mathematics, physics, engineering and even art. This book offers teachers and students an opportunity to learn about complex numbers in a state-of-the-art computational environment. The innovative approach also offers insights into many areas too often neglected in a student treatment, including complex chaos and mathematical art. Thus readers can also use the book for self-study and for enrichment. The use of Mathematica enables the author to cover several topics that are often absent from a traditional treatment. Students are also led, optionally, into cubic or quartic equations, investigations of symmetric chaos and advanced conformal mapping. A CD is included which contains a live version of the book: in particular all the Mathematica code enables the user to run computer experiments.
Cohesively Incorporates Statistical Theory with R Implementation Since the publication of the popular first edition of this comprehensive textbook, the contributed R packages on CRAN have increased from around 1,000 to over 6,000. Designed for an intermediate undergraduate course, Probability and Statistics with R, Second Edition explores how some of these new packages make analysis easier and more intuitive as well as create more visually pleasing graphs. New to the Second Edition Improvements to existing examples, problems, concepts, data, and functions New examples and exercises that use the most modern functions Coverage probability of a confidence interval and model validation Highlighted R code for calculations and graph creation Gets Students Up to Date on Practical Statistical Topics Keeping pace with today's statistical landscape, this textbook expands your students' knowledge of the practice of statistics. It effectively links statistical concepts with R procedures, empowering students to solve a vast array of real statistical problems with R. Web Resources A supplementary website offers solutions to odd exercises and templates for homework assignments while the data sets and R functions are available on CRAN.
An Introduction to Statistical Learning provides an accessible overview of the field of statistical learning, an essential toolset for making sense of the vast and complex data sets that have emerged in fields ranging from biology to finance to marketing to astrophysics in the past twenty years. This book presents some of the most important modeling and prediction techniques, along with relevant applications. Topics include linear regression, classification, resampling methods, shrinkage approaches, tree-based methods, support vector machines, clustering, deep learning, survival analysis, multiple testing, and more. Color graphics and real-world examples are used to illustrate the methods presented. Since the goal of this textbook is to facilitate the use of these statistical learning techniques by practitioners in science, industry, and other fields, each chapter contains a tutorial on implementing the analyses and methods presented in R, an extremely popular open source statistical software platform. Two of the authors co-wrote The Elements of Statistical Learning (Hastie, Tibshirani and Friedman, 2nd edition 2009), a popular reference book for statistics and machine learning researchers. An Introduction to Statistical Learning covers many of the same topics, but at a level accessible to a much broader audience. This book is targeted at statisticians and non-statisticians alike who wish to use cutting-edge statistical learning techniques to analyze their data. The text assumes only a previous course in linear regression and no knowledge of matrix algebra. This Second Edition features new chapters on deep learning, survival analysis, and multiple testing, as well as expanded treatments of naive Bayes, generalized linear models, Bayesian additive regression trees, and matrix completion. R code has been updated throughout to ensure compatibility.
In honor of professor and renowned statistician R. Dennis Cook, this festschrift explores his influential contributions to an array of statistical disciplines ranging from experimental design and population genetics, to statistical diagnostics and all areas of regression-related inference and analysis. Since the early 1990s, Prof. Cook has led the development of dimension reduction methodology in three distinct but related regression contexts: envelopes, sufficient dimension reduction (SDR), and regression graphics. In particular, he has made fundamental and pioneering contributions to SDR, inventing or co-inventing many popular dimension reduction methods, such as sliced average variance estimation, the minimum discrepancy approach, model-free variable selection, and sufficient dimension reduction subspaces. A prolific researcher and mentor, Prof. Cook is known for his ability to identify research problems in statistics that are both challenging and important, as well as his deep appreciation for the applied side of statistics. This collection of Prof. Cook's collaborators, colleagues, friends, and former students reflects the broad array of his contributions to the research and instructional arenas of statistics.
The investigation of the role of mechanical and mechano-chemical interactions in cellular processes and tissue development is a rapidly growing research field in the life sciences and in biomedical engineering. Quantitative understanding of this important area in the study of biological systems requires the development of adequate mathematical models for the simulation of the evolution of these systems in space and time. Since expertise in various fields is necessary, this calls for a multidisciplinary approach. This edited volume connects basic physical, biological, and physiological concepts to methods for the mathematical modeling of various materials by pursuing a multiscale approach, from subcellular to organ and system level. Written by active researchers, each chapter provides a detailed introduction to a given field, illustrates various approaches to creating models, and explores recent advances and future research perspectives. Topics covered include molecular dynamics simulations of lipid membranes, phenomenological continuum mechanics of tissue growth, and translational cardiovascular modeling. Modeling Biomaterials will be a valuable resource for both non-specialists and experienced researchers from various domains of science, such as applied mathematics, biophysics, computational physiology, and medicine.
Metrics are a hot topic. Executive leadership, boards of directors, management, and customers are all asking for data-based decisions. As a result, many managers, professionals, and change agents are asked to develop metrics, but have no clear idea of how to produce meaningful ones. Wouldn't it be great to have a simple explanation of how to collect, analyze, report, and use measurements to improve your organization? Metrics: How to Improve Key Business Results provides that explanation and the tools you'll need to make your organization more effective. Not only does the book explain the why of metrics, but it walks you through a step-by-step process for creating a report card that provides a clear picture of organizational health and how well you satisfy customer needs. Metrics will help you to measure the right things, the right way - the first time. No wasted effort, no chasing data. The report card provides a simple tool for viewing the health of your organization, from the outside in.You will learn how to measure the key components of the report card and thereby improve real measures of business success, like repeat customers, customer loyalty, and word-of-mouth advertising.This book: * Provides a step-by-step guide for building an organizational effectiveness report card * Takes you from identifying key services and products and using metrics, to determining business strategy * Provides examples of how to identify, collect, analyze, and report metrics that will be immediately useful for improving all aspects of the enterprise, including IT What you'll learn * Understand the difference between data, measures, information, and metrics * Identify root performance questions to ensure you build the right metrics * Develop meaningful and accurate metrics using concrete, easy-to-follow instructions * Avoid the high risks that come with collecting, analyzing, reporting, and using complex data * Formulate practical answers to data-based questions * Select and use the proper tools for creating, implementing, and using metrics * Learn one of the most powerful methods yet invented for improving organizational results Who this book is for Metrics: How to Improve Key Business Results was written for senior managers who need to improve key results.Equally, the book is for the department heads, middle managers, analysts, IT professionals, and change agents responsible for collecting, analyzing, and reporting metrics. Finally, it's for those who have to chase data and find meaningful answers to the interesting questions executives ponder. Table of Contents * Introduction: Who, What, Where, When, Why, and How You Use Metrics * Establishing a Common Language * Where to Begin: Planning a Good Metric * Using Metrics as Indicators * Using the Answer Key * Start with Effectiveness * Triangulation: Essential to Creating Effective Metrics * Expectations: How to View Data in a Meaningful Way * Creating and Interpreting the Metrics Report Card * Final Product: the Metrics Report Card * Employing Advanced Metrics * Creating the Service Catalog * Establishing Standards and Benchmarks * Respecting the Power of Metrics * Avoiding the Research Trap * Embracing Your Organization's Uniqueness * Appendix: Metrics Tools to Use and Useful Resources
This book is open access under a CC BY 4.0 license This open access book brings together the latest genome base prediction models currently being used by statisticians, breeders and data scientists. It provides an accessible way to understand the theory behind each statistical learning tool, the required pre-processing, the basics of model building, how to train statistical learning methods, the basic R scripts needed to implement each statistical learning tool, and the output of each tool. To do so, for each tool the book provides background theory, some elements of the R statistical software for its implementation, the conceptual underpinnings, and at least two illustrative examples with data from real-world genomic selection experiments. Lastly, worked-out examples help readers check their own comprehension.The book will greatly appeal to readers in plant (and animal) breeding, geneticists and statisticians, as it provides in a very accessible way the necessary theory, the appropriate R code, and illustrative examples for a complete understanding of each statistical learning tool. In addition, it weighs the advantages and disadvantages of each tool.
Inverse problems such as imaging or parameter identification deal with the recovery of unknown quantities from indirect observations, connected via a model describing the underlying context. While traditionally inverse problems are formulated and investigated in a static setting, we observe a significant increase of interest in time-dependence in a growing number of important applications over the last few years. Here, time-dependence affects a) the unknown function to be recovered and / or b) the observed data and / or c) the underlying process. Challenging applications in the field of imaging and parameter identification are techniques such as photoacoustic tomography, elastography, dynamic computerized or emission tomography, dynamic magnetic resonance imaging, super-resolution in image sequences and videos, health monitoring of elastic structures, optical flow problems or magnetic particle imaging to name only a few. Such problems demand for innovation concerning their mathematical description and analysis as well as computational approaches for their solution.
Learn how to program by diving into the R language, and then use your newfound skills to solve practical data science problems. With this book, you'll learn how to load data, assemble and disassemble data objects, navigate R's environment system, write your own functions, and use all of R's programming tools. RStudio Master Instructor Garrett Grolemund not only teaches you how to program, but also shows you how to get more from R than just visualizing and modeling data. You'll gain valuable programming skills and support your work as a data scientist at the same time. Work hands-on with three practical data analysis projects based on casino games Store, retrieve, and change data values in your computer's memory Write programs and simulations that outperform those written by typical R users Use R programming tools such as if else statements, for loops, and S3 classes Learn how to write lightning-fast vectorized R code Take advantage of R's package system and debugging tools Practice and apply R programming concepts as you learn them
Statistics and computing share many close relationships. Computing
now permeates every aspect of statistics, from pure description to
the development of statistical theory. At the same time, the
computational methods used in statistical work span much of
computer science. Elements of Statistical Computing covers the
broad usage of computing in statistics. It provides a comprehensive
account of the most important computational statistics. Included
are discussions of numerical analysis, numerical integration, and
smoothing.
This advanced textbook explores small area estimation techniques, covers the underlying mathematical and statistical theory and offers hands-on support with their implementation. It presents the theory in a rigorous way and compares and contrasts various statistical methodologies, helping readers understand how to develop new methodologies for small area estimation. It also includes numerous sample applications of small area estimation techniques. The underlying R code is provided in the text and applied to four datasets that mimic data from labor markets and living conditions surveys, where the socioeconomic indicators include the small area estimation of total unemployment, unemployment rates, average annual household incomes and poverty indicators. Given its scope, the book will be useful for master and PhD students, and for official and other applied statisticians.
The Distributed and Unified Numerics Environment (Dune) is a set of open-source C++ libraries for the implementation of finite element and finite volume methods. Over the last 15 years it has become one of the most commonly used libraries for the implementation of new, efficient simulation methods in science and engineering. Describing the main Dune libraries in detail, this book covers access to core features like grids, shape functions, and linear algebra, but also higher-level topics like function space bases and assemblers. It includes extensive information on programmer interfaces, together with a wealth of completed examples that illustrate how these interfaces are used in practice. After having read the book, readers will be prepared to write their own advanced finite element simulators, tapping the power of Dune to do so.
This book presents strategies for analyzing qualitative and mixed methods data with MAXQDA software, and provides guidance on implementing a variety of research methods and approaches, e.g. grounded theory, discourse analysis and qualitative content analysis, using the software. In addition, it explains specific topics, such as transcription, building a coding frame, visualization, analysis of videos, concept maps, group comparisons and the creation of literature reviews. The book is intended for masters and PhD students as well as researchers and practitioners dealing with qualitative data in various disciplines, including the educational and social sciences, psychology, public health, business or economics.
A Hands-On Approach to Understanding and Using Actuarial Models Computational Actuarial Science with R provides an introduction to the computational aspects of actuarial science. Using simple R code, the book helps you understand the algorithms involved in actuarial computations. It also covers more advanced topics, such as parallel computing and C/C++ embedded codes. After an introduction to the R language, the book is divided into four parts. The first one addresses methodology and statistical modeling issues. The second part discusses the computational facets of life insurance, including life contingencies calculations and prospective life tables. Focusing on finance from an actuarial perspective, the next part presents techniques for modeling stock prices, nonlinear time series, yield curves, interest rates, and portfolio optimization. The last part explains how to use R to deal with computational issues of nonlife insurance. Taking a do-it-yourself approach to understanding algorithms, this book demystifies the computational aspects of actuarial science. It shows that even complex computations can usually be done without too much trouble. Datasets used in the text are available in an R package (CASdatasets).
This volume, the third of a series, consists of applications of Mathematica (R) to a potpourri of more advanced topics. These include differential geometry of curves and surfaces, differential equations and special functions and complex analysis. Some of the newest features of Mathematica (R) are demonstrated and explained and some problems with the current implementation pointed out and possible future improvements suggested. Contains a large number of worked out examples. Explains some of the most recent mathematical features of Mathematica (R). Considers topics discussed rarely or not at all in the context of Mathematica (R). Can be used to supplement several different courses. Based on actual university courses.
This book offers a systematic and rigorous treatment of continuous-time Markov decision processes, covering both theory and possible applications to queueing systems, epidemiology, finance, and other fields. Unlike most books on the subject, much attention is paid to problems with functional constraints and the realizability of strategies. Three major methods of investigations are presented, based on dynamic programming, linear programming, and reduction to discrete-time problems. Although the main focus is on models with total (discounted or undiscounted) cost criteria, models with average cost criteria and with impulsive controls are also discussed in depth. The book is self-contained. A separate chapter is devoted to Markov pure jump processes and the appendices collect the requisite background on real analysis and applied probability. All the statements in the main text are proved in detail. Researchers and graduate students in applied probability, operational research, statistics and engineering will find this monograph interesting, useful and valuable.
Highlighting the latest advances in nonparametric and semiparametric statistics, this book gathers selected peer-reviewed contributions presented at the 4th Conference of the International Society for Nonparametric Statistics (ISNPS), held in Salerno, Italy, on June 11-15, 2018. It covers theory, methodology, applications and computational aspects, addressing topics such as nonparametric curve estimation, regression smoothing, models for time series and more generally dependent data, varying coefficient models, symmetry testing, robust estimation, and rank-based methods for factorial design. It also discusses nonparametric and permutation solutions for several different types of data, including ordinal data, spatial data, survival data and the joint modeling of both longitudinal and time-to-event data, permutation and resampling techniques, and practical applications of nonparametric statistics. The International Society for Nonparametric Statistics is a unique global organization, and its international conferences are intended to foster the exchange of ideas and the latest advances and trends among researchers from around the world and to develop and disseminate nonparametric statistics knowledge. The ISNPS 2018 conference in Salerno was organized with the support of the American Statistical Association, the Institute of Mathematical Statistics, the Bernoulli Society for Mathematical Statistics and Probability, the Journal of Nonparametric Statistics and the University of Salerno.
This book provides an introduction to quantitative marketing with Python. The book presents a hands-on approach to using Python for real marketing questions, organized by key topic areas. Following the Python scientific computing movement toward reproducible research, the book presents all analyses in Colab notebooks, which integrate code, figures, tables, and annotation in a single file. The code notebooks for each chapter may be copied, adapted, and reused in one's own analyses. The book also introduces the usage of machine learning predictive models using the Python sklearn package in the context of marketing research. This book is designed for three groups of readers: experienced marketing researchers who wish to learn to program in Python, coming from tools and languages such as R, SAS, or SPSS; analysts or students who already program in Python and wish to learn about marketing applications; and undergraduate or graduate marketing students with little or no programming background. It presumes only an introductory level of familiarity with formal statistics and contains a minimum of mathematics.
This text examines the goals of data analysis with respect to enhancing knowledge, and identifies data summarization and correlation analysis as the core issues. Data summarization, both quantitative and categorical, is treated within the encoder-decoder paradigm bringing forward a number of mathematically supported insights into the methods and relations between them. Two Chapters describe methods for categorical summarization: partitioning, divisive clustering and separate cluster finding and another explain the methods for quantitative summarization, Principal Component Analysis and PageRank. Features: * An in-depth presentation of K-means partitioning including a corresponding Pythagorean decomposition of the data scatter. * Advice regarding such issues as clustering of categorical and mixed scale data, similarity and network data, interpretation aids, anomalous clusters, the number of clusters, etc. * Thorough attention to data-driven modelling including a number of mathematically stated relations between statistical and geometrical concepts including those between goodness-of-fit criteria for decision trees and data standardization, similarity and consensus clustering, modularity clustering and uniform partitioning. New edition highlights: * Inclusion of ranking issues such as Google PageRank, linear stratification and tied rankings median, consensus clustering, semi-average clustering, one-cluster clustering * Restructured to make the logics more straightforward and sections self-contained Core Data Analysis: Summarization, Correlation and Visualization is aimed at those who are eager to participate in developing the field as well as appealing to novices and practitioners.
This book is a text for a one-semester course for upper-level undergraduates and beginning graduate students in engineering, science, and mathematics. Prerequisites are a first course in the theory of ODEs and a survey course in numerical analysis, in addition to specific programming experience, preferably in MATLAB, and knowledge of elementary matrix theory. Professionals will also find that this useful concise reference contains reviews of technical issues and realistic and detailed examples. The programs for the examples are supplied on the accompanying web site and can serve as templates for solving other problems. Each chapter begins with a discussion of the "facts of life" for the problem, mainly by means of examples. Numerical methods for the problem are then developed, but only those methods most widely used. The treatment of each method is brief and technical issues are minimized, but all the issues important in practice and for understaning the codes are discussed. The last part of each chapter is a tutorial that shows how to solve problems by means of small, but realistic, examples.
This book is dedicated to the systematization and development of models, methods, and algorithms for queuing systems with correlated arrivals. After first setting up the basic tools needed for the study of queuing theory, the authors concentrate on complicated systems: multi-server systems with phase type distribution of service time or single-server queues with arbitrary distribution of service time or semi-Markovian service. They pay special attention to practically important retrial queues, tandem queues, and queues with unreliable servers. Mathematical models of networks and queuing systems are widely used for the study and optimization of various technical, physical, economic, industrial, and administrative systems, and this book will be valuable for researchers, graduate students, and practitioners in these domains.
This thesis presents a revolutionary technique for modelling the dynamics of a quantum system that is strongly coupled to its immediate environment. This is a challenging but timely problem. In particular it is relevant for modelling decoherence in devices such as quantum information processors, and how quantum information moves between spatially separated parts of a quantum system. The key feature of this work is a novel way to represent the dynamics of general open quantum systems as tensor networks, a result which has connections with the Feynman operator calculus and process tensor approaches to quantum mechanics. The tensor network methodology developed here has proven to be extremely powerful: For many situations it may be the most efficient way of calculating open quantum dynamics. This work is abounds with new ideas and invention, and is likely to have a very significant impact on future generations of physicists.
This open access book presents a set of basic techniques for estimating the benefit of IT development projects and portfolios. It also offers methods for monitoring how much of that estimated benefit is being achieved during projects. Readers can then use these benefit estimates together with cost estimates to create a benefit/cost index to help them decide which functionalities to send into construction and in what order. This allows them to focus on constructing the functionality that offers the best value for money at an early stage. Although benefits management involves a wide range of activities in addition to estimation and monitoring, the techniques in this book provides a clear guide to achieving what has always been the goal of project and portfolio stakeholders: developing systems that produce as much usefulness and value as possible for the money invested. The techniques can also help deal with vicarious motives and obstacles that prevent this happening. The book equips readers to recognize when a project budget should not be spent in full and resources be allocated elsewhere in a portfolio instead. It also provides development managers and upper management with common ground as a basis for making informed decisions.
This book discusses the development of the Rosenbrock-Wanner methods from the origins of the idea to current research with the stable and efficient numerical solution and differential-algebraic systems of equations, still in focus. The reader gets a comprehensive insight into the classical methods as well as into the development and properties of novel W-methods, two-step and exponential Rosenbrock methods. In addition, descriptive applications from the fields of water and hydrogen network simulation and visual computing are presented.
This book provides a general introduction to Sequential Monte Carlo (SMC) methods, also known as particle filters. These methods have become a staple for the sequential analysis of data in such diverse fields as signal processing, epidemiology, machine learning, population ecology, quantitative finance, and robotics. The coverage is comprehensive, ranging from the underlying theory to computational implementation, methodology, and diverse applications in various areas of science. This is achieved by describing SMC algorithms as particular cases of a general framework, which involves concepts such as Feynman-Kac distributions, and tools such as importance sampling and resampling. This general framework is used consistently throughout the book. Extensive coverage is provided on sequential learning (filtering, smoothing) of state-space (hidden Markov) models, as this remains an important application of SMC methods. More recent applications, such as parameter estimation of these models (through e.g. particle Markov chain Monte Carlo techniques) and the simulation of challenging probability distributions (in e.g. Bayesian inference or rare-event problems), are also discussed. The book may be used either as a graduate text on Sequential Monte Carlo methods and state-space modeling, or as a general reference work on the area. Each chapter includes a set of exercises for self-study, a comprehensive bibliography, and a "Python corner," which discusses the practical implementation of the methods covered. In addition, the book comes with an open source Python library, which implements all the algorithms described in the book, and contains all the programs that were used to perform the numerical experiments. |
You may like...
Spatial Regression Analysis Using…
Daniel A. Griffith, Yongwan Chun, …
Paperback
R3,015
Discovery Miles 30 150
JMP Essentials - An Illustrated Guide…
Curt Hinrichs, Chuck Boiler, …
Hardcover
R2,063
Discovery Miles 20 630
|