![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer software packages > Other software packages > Mathematical & statistical software
This book presents theoretical modeling and numerical simulations applied to drive several applications towards Industrial Revolution 4.0 (IR 4.0). The topics discussed range from theoretical parts to extensive simulations involving many efficient algorithms as well as various statistical techniques. This book is suitable for postgraduate students, researchers as well as other scientists who are working in mathematics, statistics and numerical modeling and simulation.
This book chronicles a 10-year introduction of blended learning into the delivery at a leading technological university, with a longstanding tradition of technology-enabled teaching and learning, and state-of-the-art infrastructure. Hence, both teachers and students were familiar with the idea of online courses. Despite this, the longitudinal experiment did not proceed as expected. Though few technical problems, it required behavioural changes from teachers and learners, thus unearthing a host of socio-technical issues, challenges, and conundrums. With the undercurrent of design ideals such as "tech for good", any industrial sector must examine whether digital platforms are credible substitutes or at best complementary. In this era of Industry 4.0, higher education, like any other industry, should not be about the creative destruction of what we value in universities, but their digital transformation. The book concludes with an agenda for large, repeatable Randomised Controlled Trials (RCTs) to validate digital platforms that could fulfil the aspirations of the key stakeholder groups - students, faculty, and regulators as well as delving into the role of Massive Open Online Courses (MOOCs) as surrogates for "fees-free" higher education and whether the design of such a HiEd 4.0 platform is even a credible proposition. Specifically, the book examines the data-driven evidence within a design-based research methodology to present outcomes of two alternative instructional designs evaluated - traditional lecturing and blended learning. Based on the research findings and statistical analysis, it concludes that the inexorable shift to online delivery of education must be guided by informed educational management and innovation.
This book provides a concise point of reference for the most commonly used regression methods. It begins with linear and nonlinear regression for normally distributed data, logistic regression for binomially distributed data, and Poisson regression and negative-binomial regression for count data. It then progresses to these regression models that work with longitudinal and multi-level data structures. The volume is designed to guide the transition from classical to more advanced regression modeling, as well as to contribute to the rapid development of statistics and data science. With data and computing programs available to facilitate readers' learning experience, Statistical Regression Modeling promotes the applications of R in linear, nonlinear, longitudinal and multi-level regression. All included datasets, as well as the associated R program in packages nlme and lme4 for multi-level regression, are detailed in Appendix A. This book will be valuable in graduate courses on applied regression, as well as for practitioners and researchers in the fields of data science, statistical analytics, public health, and related fields.
Inverse problems such as imaging or parameter identification deal with the recovery of unknown quantities from indirect observations, connected via a model describing the underlying context. While traditionally inverse problems are formulated and investigated in a static setting, we observe a significant increase of interest in time-dependence in a growing number of important applications over the last few years. Here, time-dependence affects a) the unknown function to be recovered and / or b) the observed data and / or c) the underlying process. Challenging applications in the field of imaging and parameter identification are techniques such as photoacoustic tomography, elastography, dynamic computerized or emission tomography, dynamic magnetic resonance imaging, super-resolution in image sequences and videos, health monitoring of elastic structures, optical flow problems or magnetic particle imaging to name only a few. Such problems demand for innovation concerning their mathematical description and analysis as well as computational approaches for their solution.
This book offers a systematic and rigorous treatment of continuous-time Markov decision processes, covering both theory and possible applications to queueing systems, epidemiology, finance, and other fields. Unlike most books on the subject, much attention is paid to problems with functional constraints and the realizability of strategies. Three major methods of investigations are presented, based on dynamic programming, linear programming, and reduction to discrete-time problems. Although the main focus is on models with total (discounted or undiscounted) cost criteria, models with average cost criteria and with impulsive controls are also discussed in depth. The book is self-contained. A separate chapter is devoted to Markov pure jump processes and the appendices collect the requisite background on real analysis and applied probability. All the statements in the main text are proved in detail. Researchers and graduate students in applied probability, operational research, statistics and engineering will find this monograph interesting, useful and valuable.
This book provides an accessible introduction and practical guidelines to apply asymmetric multidimensional scaling, cluster analysis, and related methods to asymmetric one-mode two-way and three-way asymmetric data. A major objective of this book is to present to applied researchers a set of methods and algorithms for graphical representation and clustering of asymmetric relationships. Data frequently concern measurements of asymmetric relationships between pairs of objects from a given set (e.g., subjects, variables, attributes,...), collected in one or more matrices. Examples abound in many different fields such as psychology, sociology, marketing research, and linguistics and more recently several applications have appeared in technological areas including cybernetics, air traffic control, robotics, and network analysis. The capabilities of the presented algorithms are illustrated by carefully chosen examples and supported by extensive data analyses. A review of the specialized statistical software available for the applications is also provided. This monograph is highly recommended to readers who need a complete and up-to-date reference on methods for asymmetric proximity data analysis.
The nonequilibrium behavior of nanoscopic and biological systems, which are typically strongly fluctuating, is a major focus of current research. Lately, much progress has been made in understanding such systems from a thermodynamic perspective. However, new theoretical challenges emerge when the fluctuating system is additionally subject to time delay, e.g. due to the presence of feedback loops. This thesis advances this young and vibrant research field in several directions. The first main contribution concerns the probabilistic description of time-delayed systems; e.g. by introducing a versatile approximation scheme for nonlinear delay systems. Second, it reveals that delay can induce intriguing thermodynamic properties such as anomalous (reversed) heat flow. More generally, the thesis shows how to treat the thermodynamics of non-Markovian systems by introducing auxiliary variables. It turns out that delayed feedback is inextricably linked to nonreciprocal coupling, information flow, and to net energy input on the fluctuating level.
An Introduction to Stata for Health Researchers, Fifth Edition updates this classic book that has become a standard reference for health researchers. As with previous editions, readers will learn to work effectively in Stata to perform data management, compute descriptive statistics, create meaningful graphs, fit regression models, and perform survival analysis. The fifth edition adds examples of performing power, precision, and sample-size analysis; working with Unicode characters; managing data with ICD-9 and ICD-10 codes; and creating customized tables. With many worked examples and downloadable datasets, this text is the ideal resource for hands-on learning, whether for students in a statistics course or for researchers in fields such as epidemiology, biostatistics, and public health who are learning to use Stata's tools for health research.
This thesis presents a revolutionary technique for modelling the dynamics of a quantum system that is strongly coupled to its immediate environment. This is a challenging but timely problem. In particular it is relevant for modelling decoherence in devices such as quantum information processors, and how quantum information moves between spatially separated parts of a quantum system. The key feature of this work is a novel way to represent the dynamics of general open quantum systems as tensor networks, a result which has connections with the Feynman operator calculus and process tensor approaches to quantum mechanics. The tensor network methodology developed here has proven to be extremely powerful: For many situations it may be the most efficient way of calculating open quantum dynamics. This work is abounds with new ideas and invention, and is likely to have a very significant impact on future generations of physicists.
Mathematical Models of Biological Systems provides a practical introduction to basic mathematical modelling methodology and analysis. It covers a variety of biological applications and uses these topics in turn to highlight key components in the art of modelling. Its primary aim is to give students the tools to translate simple, real-world biological problems into rigorous mathematical models. A secondary aim is to teach the reader how to critically assess the modelling components in the primary life science literature. The book covers deterministic as well as stochastic dynamics, continuous-time as well as discrete-time dynamics, partial differential equations, dimensional analysis, and curve fitting/parameter estimation. It contains numerous case studies, graded from elementary examples to more complicated problems, as well as a general treatment of good modelling practice. Although the book assumes a basic background in mathematics, specifically beginning calculus and elementary statistics, all requisite material is included in a series of appendices.
This book gathers a selection of invited and contributed lectures from the European Conference on Numerical Mathematics and Advanced Applications (ENUMATH) held in Lausanne, Switzerland, August 26-30, 2013. It provides an overview of recent developments in numerical analysis, computational mathematics and applications from leading experts in the field. New results on finite element methods, multiscale methods, numerical linear algebra and discretization techniques for fluid mechanics and optics are presented. As such, the book offers a valuable resource for a wide range of readers looking for a state-of-the-art overview of advanced techniques, algorithms and results in numerical mathematics and scientific computing.
This book shows how information theory, probability, statistics, mathematics and personal computers can be applied to the exploration of numbers and proportions in music. It brings the methods of scientific and quantitative thinking to questions like: What are the ways of encoding a message in music and how can we be sure of the correct decoding? How do claims of names hidden in the notes of a score stand up to scientific analysis? How many ways are there of obtaining proportions and are they due to chance? After thoroughly exploring the ways of encoding information in music, the ambiguities of numerical alphabets and the words to be found "hidden" in a score, the book presents a novel way of exploring the proportions in a composition with a purpose-built computer program and gives example results from the application of the techniques. These include information theory, combinatorics, probability, hypothesis testing, Monte Carlo simulation and Bayesian networks, presented in an easily understandable form including their development from ancient history through the life and times of J. S. Bach, making connections between science, philosophy, art, architecture, particle physics, calculating machines and artificial intelligence. For the practitioner the book points out the pitfalls of various psychological fallacies and biases and includes succinct points of guidance for anyone involved in this type of research. This book will be useful to anyone who intends to use a scientific approach to the humanities, particularly music, and will appeal to anyone who is interested in the intersection between the arts and science.With a foreword by Ruth Tatlow (Uppsala University), award winning author of Bach's Numbers: Compositional Proportion and Significance and Bach and the Riddle of the Number Alphabet."With this study Alan Shepherd opens a much-needed examination of the wide range of mathematical claims that have been made about J. S. Bach's music, offering both tools and methodological cautions with the potential to help clarify old problems." Daniel R. Melamed, Professor of Music in Musicology, Indiana University
This book introduces readers to various signal processing models that have been used in analyzing periodic data, and discusses the statistical and computational methods involved. Signal processing can broadly be considered to be the recovery of information from physical observations. The received signals are usually disturbed by thermal, electrical, atmospheric or intentional interferences, and due to their random nature, statistical techniques play an important role in their analysis. Statistics is also used in the formulation of appropriate models to describe the behavior of systems, the development of appropriate techniques for estimation of model parameters and the assessment of the model performances. Analyzing different real-world data sets to illustrate how different models can be used in practice, and highlighting open problems for future research, the book is a valuable resource for senior undergraduate and graduate students specializing in mathematics or statistics.
A Tour of Data Science: Learn R and Python in Parallel covers the fundamentals of data science, including programming, statistics, optimization, and machine learning in a single short book. It does not cover everything, but rather, teaches the key concepts and topics in Data Science. It also covers two of the most popular programming languages used in Data Science, R and Python, in one source. Key features: Allows you to learn R and Python in parallel Cover statistics, programming, optimization and predictive modelling, and the popular data manipulation tools - data.table and pandas Provides a concise and accessible presentation Includes machine learning algorithms implemented from scratch, linear regression, lasso, ridge, logistic regression, gradient boosting trees, etc. Appealing to data scientists, statisticians, quantitative analysts, and others who want to learn programming with R and Python from a data science perspective.
The book "Analysis and Design of Control Systems using MATLAB", is designed as a supplement to an introductory course in feedback control systems for undergraduate or graduate engineering students of all disciplines. Feedback control systems engineering is a multidisciplinary subject and presents a control engineering methodology based on mathematical fundamentals and stresses physical system modeling.This book includes the coverage of classical methods of control systems engineering: introduction to control systems, matrix analysis, Laplace transforms, mathematical modeling of dynamic systems, control system representation, performance and stability of feedback systems, analysis and design of feedback control systems, state space analysis and design, and MATLAB basics and MATLAB tutorial. The numerous worked examples offer detailed explanations, and guide the students through each set of problems to enable them to save a great deal of time and effort in arriving at an understanding of problems in this subject. Extensive references to guide the students to further sources of information on control systems and MATLAB is provided. In addition to students, practising engineers will also find this book immensely useful.
This book highlights recent advances in natural computing, including biology and its theory, bio-inspired computing, computational aesthetics, computational models and theories, computing with natural media, philosophy of natural computing and educational technology. It presents extended versions of the best papers selected from the symposium "7th International Workshop on Natural Computing" (IWNC7), held in Tokyo, Japan, in 2013. The target audience is not limited to researchers working in natural computing but also those active in biological engineering, fine/media art design, aesthetics and philosophy.
For courses on SPSS. SPSS is, essentially, a visually-driven program, but most texts rely primarily on a verbal approach to describe its use. A Visual Approach to SPSS for Windows is the first text of its kind to employ what the author refers to as "visual sequencing" to teach students how to use SPSS.
This book covers all the topics found in introductory descriptive statistics courses, including simple linear regression and time series analysis, the fundamentals of inferential statistics (probability theory, random sampling and estimation theory), and inferential statistics itself (confidence intervals, testing). Each chapter starts with the necessary theoretical background, which is followed by a variety of examples. The core examples are based on the content of the respective chapter, while the advanced examples, designed to deepen students' knowledge, also draw on information and material from previous chapters. The enhanced online version helps students grasp the complexity and the practical relevance of statistical analysis through interactive examples and is suitable for undergraduate and graduate students taking their first statistics courses, as well as for undergraduate students in non-mathematical fields, e.g. economics, the social sciences etc.
Understanding Regression Analysis unifies diverse regression applications including the classical model, ANOVA models, generalized models including Poisson, Negative binomial, logistic, and survival, neural networks, and decision trees under a common umbrella -- namely, the conditional distribution model. It explains why the conditional distribution model is the correct model, and it also explains (proves) why the assumptions of the classical regression model are wrong. Unlike other regression books, this one from the outset takes a realistic approach that all models are just approximations. Hence, the emphasis is to model Nature's processes realistically, rather than to assume (incorrectly) that Nature works in particular, constrained ways. Key features of the book include: Numerous worked examples using the R software Key points and self-study questions displayed "just-in-time" within chapters Simple mathematical explanations ("baby proofs") of key concepts Clear explanations and applications of statistical significance (p-values), incorporating the American Statistical Association guidelines Use of "data-generating process" terminology rather than "population" Random-X framework is assumed throughout (the fixed-X case is presented as a special case of the random-X case) Clear explanations of probabilistic modelling, including likelihood-based methods Use of simulations throughout to explain concepts and to perform data analyses This book has a strong orientation towards science in general, as well as chapter-review and self-study questions, so it can be used as a textbook for research-oriented students in the social, biological and medical, and physical and engineering sciences. As well, its mathematical emphasis makes it ideal for a text in mathematics and statistics courses. With its numerous worked examples, it is also ideally suited to be a reference book for all scientists.
- includes MATLABr fundamentals, matrices, arrays, general graphics and specialized plots in quality assurance problems, script files, ordinary and partial differential equations - gives calculation of six sigma, total quality management, time series forecasting, reliability, process improvement, metrology, quality control and assurance, measurement and testing techniques - provides tools for graphical presentation, basic and special statistics and testing, ordinary and partial differential solvers, and fitting tools - includes comprehensive command information in tables Many books are available on MATLABr programming for engineers in general or in some specific area, but none in the highly topical field of quality assurance (QA). MATLABr in quality assurance sciences fills this gap as a compact guide for students, engineers, and scientists in this field. It concentrates on MATLABr fundamentals with examples of application to a wide range of current problems from general, nano and bio-technology, and statistical control, to medicine and industrial management. Examples cover both the school and advanced level; comprising calculations of total quality management, six sigma, time series, process improvement, metrology, quality control, human factors in quality assurance, measurement and testing techniques, quality project and function management, and customer satisfaction. The book covers key topics, including: the basics of software with examples; graphics and representations; numerical computation, scripts and functions for QA calculations; ODE and PDEPE solvers applied to QA problems; curve fitting and time series tool interfaces in calculations of quality; and statistics calculations applied to quality testing.
Applied Statistics for Environmental Science with R presents the theory and application of statistical techniques in environmental science and aids researchers in choosing the appropriate statistical technique for analyzing their data. Focusing on the use of univariate and multivariate statistical methods, this book acts as a step-by-step resource to facilitate understanding in the use of R statistical software for interpreting data in the field of environmental science. Researchers utilizing statistical analysis in environmental science and engineering will find this book to be essential in solving their day-to-day research problems.
This book provides a reference for people working in the design, development, and manufacturing of medical devices. While there are no statistical methods specifically intended for medical devices, there are methods that are commonly applied to various problems in the design, manufacturing, and quality control of medical devices. The aim of this book is not to turn everyone working in the medical device industries into mathematical statisticians; rather, the goal is to provide some help in thinking statistically, and knowing where to go to answer some fundamental questions, such as justifying a method used to qualify/validate equipment, or what information is necessary to support the choice of sample sizes. While, there are no statistical methods specifically designed for analysis of medical device data, there are some methods that seem to appear regularly in relation to medical devices. For example, the assessment of receiver operating characteristic curves is fundamental to development of diagnostic tests, and accelerated life testing is often critical for assessing the shelf life of medical device products. Another example is sensitivity/specificity computations are necessary for in-vitro diagnostics, and Taguchi methods can be very useful for designing devices. Even notions of equivalence and noninferiority have different interpretations in the medical device field compared to pharmacokinetics. It contains topics such as dynamic modeling, machine learning methods, equivalence testing, and experimental design, for example. This book is for those with no statistical experience, as well as those with statistical knowledgeable—with the hope to provide some insight into what methods are likely to help provide rationale for choices relating to data gathering and analysis activities for medical devices.
This book provides insights into important new developments in the area of statistical quality control and critically discusses methods used in on-line and off-line statistical quality control. The book is divided into three parts: Part I covers statistical process control, Part II deals with design of experiments, while Part III focuses on fields such as reliability theory and data quality. The 12th International Workshop on Intelligent Statistical Quality Control (Hamburg, Germany, August 16 - 19, 2016) was jointly organized by Professors Sven Knoth and Wolfgang Schmid. The contributions presented in this volume were carefully selected and reviewed by the conference's scientific program committee. Taken together, they bridge the gap between theory and practice, making the book of interest to both practitioners and researchers in the field of quality control. |
You may like...
System Assurances - Modeling and…
Prashant Johri, Adarsh Anand, …
Paperback
R2,610
Discovery Miles 26 100
Spatial Regression Analysis Using…
Daniel A. Griffith, Yongwan Chun, …
Paperback
R3,015
Discovery Miles 30 150
Essential Java for Scientists and…
Brian Hahn, Katherine Malan
Paperback
R1,266
Discovery Miles 12 660
An Introduction to Creating Standardized…
Todd Case, Yuting Tian
Hardcover
R1,501
Discovery Miles 15 010
|