![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer software packages > Other software packages > Mathematical & statistical software
Economists are regularly confronted with results of quantitative economics research. Econometrics: Theory and Applications with EViews provides a broad introduction to quantitative economic methods, for example how models arise, their underlying assumptions and how estimates of parameters or other economic quantities are computed. The author combines econometric theory with practice by demonstrating its use with the software package EViews through extensive use of screen shots. The emphasis is on understanding how to select the right method of analysis for a given situation, and how to actually apply the theoretical methodology correctly. The EViews software package is available from 'Quantitive Micro Software'. Written for any undergraduate or postgraduate course in Econometrics.
A comprehensive guide to automated statistical data cleaning The production of clean data is a complex and time-consuming process that requires both technical know-how and statistical expertise. Statistical Data Cleaning brings together a wide range of techniques for cleaning textual, numeric or categorical data. This book examines technical data cleaning methods relating to data representation and data structure. A prominent role is given to statistical data validation, data cleaning based on predefined restrictions, and data cleaning strategy. Key features: Focuses on the automation of data cleaning methods, including both theory and applications written in R. Enables the reader to design data cleaning processes for either one-off analytical purposes or for setting up production systems that clean data on a regular basis. Explores statistical techniques for solving issues such as incompleteness, contradictions and outliers, integration of data cleaning components and quality monitoring. Supported by an accompanying website featuring data and R code. This book enables data scientists and statistical analysts working with data to deepen their understanding of data cleaning as well as to upgrade their practical data cleaning skills. It can also be used as material for a course in data cleaning and analyses.
A new paradigm for scientific discovery through computational tools now permeates every aspect of astronomical research. Computational astrophysics combines modern computational methods, novel hardware designs, advanced algorithms, original software implementations, and associated technologies to discover new phenomena and to make predictions in astronomy. The proceedings of IAU Symposium 362 summarizes ongoing developments in computational astrophysics through astronomers in diverse fields sharing their knowledge and approaches. It focuses on computational methods applied to speed up and broaden the scope of scientific studies, such as finding trends in observational data, high performance computing, automated search algorithms, and model predictability. Experts discuss a palette of challenging informational and technical developments, with the goal of coordinating their efforts and the improvement of techniques in pursuit of a wide range of astronomical studies, including fluid dynamics in star and galaxy evolution, exoplanets, gravitational waves, numerical relativity, data mining, and much more.
If you want to learn to use R for data analysis but aren't sure how to get started, this practical book will help you find the right path through your data. Drawing on real-world data to show you how to use different techniques in practice, it helps you progress your programming and statistics knowledge so you can apply the most appropriate tools in your research. It starts with descriptive statistics and moves through regression to advanced techniques such as structural equation modelling and Bayesian statistics, all with digestible mathematical detail for beginner researchers. The book: Shows you how to use R packages and apply functions, adjusting them to suit different datasets. Gives you the tools to try new statistical techniques and empowers you to become confident using them. Encourages you to learn by doing when running and adapting the authors' own code. Equips you with solutions to overcome the potential challenges of working with real data that may be messy or imperfect. Accompanied by online resources including screencast tutorials of R that give you step by step guidance and R scripts and datasets for you to practice with, this book is a perfect companion for any student of applied statistics or quantitative research methods courses.
"Bayesian Networks and Influence Diagrams: A Guide to Construction and Analysis, Second Edition, "provides a comprehensive guide for practitioners who wish to understand, construct, and analyze intelligent systems for decision support based on probabilistic networks. This new edition contains six new sections, in addition to fully-updated examples, tables, figures, and a revised appendix. Intended primarily for practitioners, this book does not require sophisticated mathematical skills or deep understanding of the underlying theory and methods nor does it discuss alternative technologies for reasoning under uncertainty. The theory and methods presented are illustrated through more than 140 examples, and exercises are included for the reader to check his or her level of understanding. The techniques and methods presented for knowledge elicitation, model construction and verification, modeling techniques and tricks, learning models from data, and analyses of models have all been developed and refined on the basis of numerous courses that the authors have held for practitioners worldwide. "
An easy-to-understand course book, based on the authentic lectures and detailed research, conducted by the authors themselves, on information optics, holography and MATLAB. This book is the first to highlight the incoherent optical system, provide up-to-date, novel digital holography techniques, and demonstrate MATLAB codes to accomplish tasks such as optical image processing and pattern recognition. This title is a comprehensive introduction to the basics of Fourier optics as well as optical image processing and digital holography. A step-by-step guide which details the vast majority of the derivations, without omitting essential steps, to facilitate a clear mathematical understanding. This book also features exercises at the end of each chapter, providing hands-on experience and consolidating understanding. An ideal companion for graduates and researchers involved in engineering and applied physics, as well as interested in the growing field of information optics.
The bestselling beginner Arduino guide, updated with new projects! Exploring Arduino makes electrical engineering and embedded software accessible. Learn step by step everything you need to know about electrical engineering, programming, and human-computer interaction through a series of increasingly complex projects. Arduino guru Jeremy Blum walks you through each build, providing code snippets and schematics that will remain useful for future projects. Projects are accompanied by downloadable source code, tips and tricks, and video tutorials to help you master Arduino. You'll gain the skills you need to develop your own microcontroller projects! This new 2nd edition has been updated to cover the rapidly-expanding Arduino ecosystem, and includes new full-color graphics for easier reference. Servo motors and stepper motors are covered in richer detail, and you'll find more excerpts about technical details behind the topics covered in the book. Wireless connectivity and the Internet-of-Things are now more prominently featured in the advanced projects to reflect Arduino's growing capabilities. You'll learn how Arduino compares to its competition, and how to determine which board is right for your project. If you're ready to start creating, this book is your ultimate guide! * Get up to date on the evolving Arduino hardware, software, and capabilities * Build projects that interface with other devices wirelessly! * Learn the basics of electrical engineering and programming * Access downloadable materials and source code for every project Whether you're a first-timer just starting out in electronics, or a pro looking to mock-up more complex builds, Arduino is a fantastic tool for building a variety of devices. This book offers a comprehensive tour of the hardware itself, plus in-depth introduction to the various peripherals, tools, and techniques used to turn your little Arduino device into something useful, artistic, and educational. Exploring Arduino is your roadmap to adventure start your journey today!
Until recently, acquiring a background in the basic methodological principles that apply to most types of investigations meant struggling to obtain results through laborious calculations. The advent of statistical software packages has removed much of the tedium and many of the errors of manual calculations and allowed a marked increase in the depth and sophistication of analyses. Although most statistics classes now incorporate some instruction in using a statistics package, most introductory texts do not. Quantitative Investigations in the Biosciences using MINITAB fills this void by providing an introduction to investigative methods that, in addition to outlining statistical principles and describing methods of calculations, also presents essential commands and interprets output from the statistics package MINITAB. The author introduces the three basic elements of investigations-design, analysis, and reporting-using an extremely accessible approach that keeps mathematical detail to a minimum. He groups statistical tests according to the type of problem they are used to examine, such as comparisons, sequential relationships, and associations. Quantitative Investigations in the Biosciences using MINITAB draws techniques and examples from a variety of subjects, ranging from physiology and biochemistry through to ecology, behavioral sciences, medicine, agriculture and horticulture, and complements the mathematical results with formal conclusions for all of the worked examples. It thus provides an ideal handbook for anyone in virtually any field who wants to apply statistical techniques to their investigations.
Written for the practicing electronics professional, Tolerance Analysis of Electronic Circuits Using MATHCADä offers a comprehensive, step-by-step treatment of methods used to perform analyses essential to the design process of circuit cards and systems of cards, including:
This compact course is written for the mathematically literate reader who wants to learn to analyze data in a principled fashion. The language of mathematics enables clear exposition that can go quite deep, quite quickly, and naturally supports an axiomatic and inductive approach to data analysis. Starting with a good grounding in probability, the reader moves to statistical inference via topics of great practical importance - simulation and sampling, as well as experimental design and data collection - that are typically displaced from introductory accounts. The core of the book then covers both standard methods and such advanced topics as multiple testing, meta-analysis, and causal inference.
Written specifically for those with no prior programming experience and minimal quantitative training, this accessible text walks behavioral science students and researchers through the process of programming using MATLAB. The book explores examples, terms, and programming needs relevant to those in the behavioral sciences and helps readers perform virtually any computational function in solving their research problems. Principles are illustrated with usable code. Each chapter opens with a list of objectives followed by new commands required to accomplish those goals. These objectives also serve as a reference to help readers easily relocate a section of interest. Sample code and output and chapter problems demonstrate how to write a program and explore a model so readers can see the results obtained using different equations and values. A web site provides solutions to selected problems and the book's program code output and examples so readers can manipulate them as needed. The outputs on the website have color, motion, and sound. Highlights of the new edition include: *Updated to reflect changes in the most recent version of MATLAB, including special tricks and new functions. *More information on debugging and common errors and more basic problems in the rudiments of MATLAB to help novice users get up and running more quickly. *A new chapter on Psychtoolbox, a suite of programs specifically geared to behavioral science research. *A new chapter on Graphical User Interfaces (GUIs) for user-friendly communication. *Increased emphasis on pre-allocation of memory, recursion, handles, and matrix algebra operators. The book opens with an overview of what is to come and tips on how to write clear programs followed by pointers for interacting with MATLAB, including its commands and how to read error messages. The matrices chapter reviews how to store and access data. Chapter 4 examines how to carry out calculations followed by a review of how to perform various actions depending on the conditions. The chapter on input and output demonstrates how to design programs to create dialogs with users (e.g., participants in studies) and read and write data to and from external files. Chapter 7 reviews the data types available in MATLAB. Readers learn how to write a program as a stand-alone module in Chapter 8. In Chapters 9 and 10 readers learn how to create line and bar graphs or reshape images. Readers learn how to create animations and sounds in Chapter 11. The book concludes with tips on how to use MATLAB with applications such as GUIs and Psychtoolbox. Intended as a primary text for Matlab courses for advanced undergraduate and/or graduate students in experimental and cognitive psychology and/or neuroscience as well as a supplementary text for labs in data (statistical) analysis, research methods, and computational modeling (programming), the book also appeals to individual researchers in these disciplines who wish to get up and running in MATLAB.
This book presents a theoretical and practical overview of computational modeling in bioengineering, focusing on a range of applications including electrical stimulation of neural and cardiac tissue, implantable drug delivery, cancer therapy, biomechanics, cardiovascular dynamics, as well as fluid-structure interaction for modelling of organs, tissues, cells and devices. It covers the basic principles of modeling and simulation with ordinary and partial differential equations using MATLAB and COMSOL Multiphysics numerical software. The target audience primarily comprises postgraduate students and researchers, but the book may also be beneficial for practitioners in the medical device industry.
CUDA is now the dominant language used for programming GPUs, one of the most exciting hardware developments of recent decades. With CUDA, you can use a desktop PC for work that would have previously required a large cluster of PCs or access to a HPC facility. As a result, CUDA is increasingly important in scientific and technical computing across the whole STEM community, from medical physics and financial modelling to big data applications and beyond. This unique book on CUDA draws on the author's passion for and long experience of developing and using computers to acquire and analyse scientific data. The result is an innovative text featuring a much richer set of examples than found in any other comparable book on GPU computing. Much attention has been paid to the C++ coding style, which is compact, elegant and efficient. A code base of examples and supporting material is available online, which readers can build on for their own projects.
R is now the most widely used statistical package/language in university statistics departments and many research organisations. Its great advantages are that for many years it has been the leading-edge statistical package/language and that it can be freely downloaded from the R web site. Its cooperative development and open code also attracts many contributors meaning that the modelling and data analysis possibilities in R are much richer than in GLIM4, and so the R edition can be substantially more comprehensive than the GLIM4 edition. This text provides a comprehensive treatment of the theory of statistical modelling in R with an emphasis on applications to practical problems and an expanded discussion of statistical theory. A wide range of case studies is provided, using the normal, binomial, Poisson, multinomial, gamma, exponential and Weibull distributions, making this book ideal for graduates and research students in applied statistics and a wide range of quantitative disciplines.
Stress Testing and Risk Integration in Banks provides a comprehensive view of the risk management activity by means of the stress testing process. An introduction to multivariate time series modeling paves the way to scenario analysis in order to assess a bank resilience against adverse macroeconomic conditions. Assets and liabilities are jointly studied to highlight the key issues that a risk manager needs to face. A multi-national bank prototype is used all over the book for diving into market, credit, and operational stress testing. Interest rate, liquidity and other major risks are also studied together with the former to outline how to implement a fully integrated risk management toolkit. Examples, business cases, and exercises worked in Matlab and R facilitate readers to develop their own models and methodologies.
This primer has been designed as a self-instructional text which serves to introduce the reader to both the principles of statistical modelling of covariance structures and to the use of the EQS software package. It is divided into three parts - the first covering the basic ideas and language of covariance structure modelling together with an introduction to the EQS package. The second section covers a wide variety of models suitable for cross-sectional and longitudinal data and the final section discusses a wide variety of practical problems. This book should be of interest to researchers in psychology, sociology and medicine who use the EQS software; applied and consultant statisticians.
This book is a valuable read for a diverse group of researchers and practitioners who analyze assessment data and construct test instruments. It focuses on the use of classical test theory (CTT) and item response theory (IRT), which are often required in the fields of psychology (e.g. for measuring psychological traits), health (e.g. for measuring the severity of disorders), and education (e.g. for measuring student performance), and makes these analytical tools accessible to a broader audience. Having taught assessment subjects to students from diverse backgrounds for a number of years, the three authors have a wealth of experience in presenting educational measurement topics, in-depth concepts and applications in an accessible format. As such, the book addresses the needs of readers who use CTT and IRT in their work but do not necessarily have an extensive mathematical background. The book also sheds light on common misconceptions in applying measurement models, and presents an integrated approach to different measurement methods, such as contrasting CTT with IRT and multidimensional IRT models with unidimensional IRT models. Wherever possible, comparisons between models are explicitly made. In addition, the book discusses concepts for test equating and differential item functioning, as well as Bayesian IRT models and plausible values using simple examples. This book can serve as a textbook for introductory courses on educational measurement, as supplementary reading for advanced courses, or as a valuable reference guide for researchers interested in analyzing student assessment data.
Statistical Programming in SAS Second Edition provides a foundation for programming to implement statistical solutions using SAS, a system that has been used to solve data analytic problems for more than 40 years. The author includes motivating examples to inspire readers to generate programming solutions. Upper-level undergraduates, beginning graduate students, and professionals involved in generating programming solutions for data-analytic problems will benefit from this book. The ideal background for a reader is some background in regression modeling and introductory experience with computer programming. The coverage of statistical programming in the second edition includes Getting data into the SAS system, engineering new features, and formatting variables Writing readable and well-documented code Structuring, implementing, and debugging programs that are well documented Creating solutions to novel problems Combining data sources, extracting parts of data sets, and reshaping data sets as needed for other analyses Generating general solutions using macros Customizing output Producing insight-inspiring data visualizations Parsing, processing, and analyzing text Programming solutions using matrices and connecting to R Processing text Programming with matrices Connecting SAS with R Covering topics that are part of both base and certification exams.
Provides researchers with a reproducible research workflow for using R/RStudio to make the entire researchprocess reproducible; from data gathering, to analysis, to presentation Includes instructions not only for creating reproducible research in R, but also extensively discusses how to take advantage of recent developments in RStudio. Emphasizes the presentation of reproducible research with non-print formats such as HTML5 slideshows, blogs, and other web-based content. Covers a range of techniques to organize and remotely store files at all stages of the research process. These techniques both streamline the research process, especially by making revisions easier, and enhance The book itself will be reproducible, as all of the data, analysis, and markup files will be made available online.
Arguably the strongest addition to numerical finance of the past decade, Algorithmic Adjoint Differentiation (AAD) is the technology implemented in modern financial software to produce thousands of accurate risk sensitivities, within seconds, on light hardware. AAD recently became a centerpiece of modern financial systems and a key skill for all quantitative analysts, developers, risk professionals or anyone involved with derivatives. It is increasingly taught in Masters and PhD programs in finance. Danske Bank's wide scale implementation of AAD in its production and regulatory systems won the In-House System of the Year 2015 Risk award. The Modern Computational Finance books, written by three of the very people who designed Danske Bank's systems, offer a unique insight into the modern implementation of financial models. The volumes combine financial modelling, mathematics and programming to resolve real life financial problems and produce effective derivatives software. This volume is a complete, self-contained learning reference for AAD, and its application in finance. AAD is explained in deep detail throughout chapters that gently lead readers from the theoretical foundations to the most delicate areas of an efficient implementation, such as memory management, parallel implementation and acceleration with expression templates. The book comes with professional source code in C++, including an efficient, up to date implementation of AAD and a generic parallel simulation library. Modern C++, high performance parallel programming and interfacing C++ with Excel are also covered. The book builds the code step-by-step, while the code illustrates the concepts and notions developed in the book.
The SPSS Survival Manual throws a lifeline to students and researchers grappling with this powerful data analysis software. In her bestselling guide, Julie Pallant takes you through the entire research process, helping you choose the right data analysis technique for your project. This edition has been updated to include up to SPSS version 26. From the formulation of research questions, to the design of the study and analysis of data, to reporting the results, Julie discusses basic and advanced statistical techniques. She outlines each technique clearly, with step-by-step procedures for performing the analysis, a detailed guide to interpreting data output and an example of how to present the results in a report. For both beginners and experienced users in Psychology, Sociology, Health Sciences, Medicine, Education, Business and related disciplines, the SPSS Survival Manual is an essential text. It is illustrated throughout with screen grabs, examples of output and tips, and is also further supported by a website with sample data and guidelines on report writing. This seventh edition is fully revised and updated to accommodate changes to IBM SPSS procedures.
This volume, the third of a series, consists of applications of Mathematica (R) to a potpourri of more advanced topics. These include differential geometry of curves and surfaces, differential equations and special functions and complex analysis. Some of the newest features of Mathematica (R) are demonstrated and explained and some problems with the current implementation pointed out and possible future improvements suggested. Contains a large number of worked out examples. Explains some of the most recent mathematical features of Mathematica (R). Considers topics discussed rarely or not at all in the context of Mathematica (R). Can be used to supplement several different courses. Based on actual university courses.
Advanced R helps you understand how R works at a fundamental level. It is designed for R programmers who want to deepen their understanding of the language, and programmers experienced in other languages who want to understand what makes R different and special. This book will teach you the foundations of R; three fundamental programming paradigms (functional, object-oriented, and metaprogramming); and powerful techniques for debugging and optimising your code. By reading this book, you will learn: The difference between an object and its name, and why the distinction is important The important vector data structures, how they fit together, and how you can pull them apart using subsetting The fine details of functions and environments The condition system, which powers messages, warnings, and errors The powerful functional programming paradigm, which can replace many for loops The three most important OO systems: S3, S4, and R6 The tidy eval toolkit for metaprogramming, which allows you to manipulate code and control evaluation Effective debugging techniques that you can deploy, regardless of how your code is run How to find and remove performance bottlenecks The second edition is a comprehensive update: New foundational chapters: "Names and values," "Control flow," and "Conditions" comprehensive coverage of object oriented programming with chapters on S3, S4, R6, and how to choose between them Much deeper coverage of metaprogramming, including the new tidy evaluation framework use of new package like rlang (http://rlang.r-lib.org), which provides a clean interface to low-level operations, and purr (http://purrr.tidyverse.org/) for functional programming Use of color in code chunks and figures Hadley Wickham is Chief Scientist at RStudio, an Adjunct Professor at Stanford University and the University of Auckland, and a member of the R Foundation. He is the lead developer of the tidyverse, a collection of R packages, including ggplot2 and dplyr, designed to support data science. He is also the author of R for Data Science (with Garrett Grolemund), R Packages, and ggplot2: Elegant Graphics for Data Analysis.
This Element is intended for students and practitioners as a gentle and intuitive introduction to the field of discrete-time yield curve modelling. I strive to be as comprehensive as possible, while still adhering to the overall premise of putting a strong focus on practical applications. In addition to a thorough description of the Nelson-Siegel family of model, the Element contains a section on the intuitive relationship between P and Q measures, one on how the structure of a Nelson-Siegel model can be retained in the arbitrage-free framework, and a dedicated section that provides a detailed explanation for the Joslin, Singleton, and Zhu (2011) model.
To write an accomplished program in the DATA step of SAS (R), programmers must understand programming logic and know how to implement and even create their own programming algorithm. Handbook of SAS (R) DATA Step Programming shows readers how best to manage and manipulate data by using the DATA step. The book helps novices avoid common mistakes resulting from a lack of understanding fundamental and unique SAS programming concepts. It explains that learning syntax does not solve all problems; rather, a thorough comprehension of SAS processing is needed for successful programming. The author also guides readers through a programming task. In most of the examples, the author first presents strategies and steps for solving the problem, then offers a solution, and finally gives a more detailed explanation of the solution. Understanding the DATA steps, particularly the program data vector (PDV), is critical to proper data manipulation and management in SAS. This book helps SAS programmers thoroughly grasp the concept of DATA step processing and write accurate programs in the DATA step. Numerous supporting materials, including data sets and programs used in the text, are available on the book's CRC Press web page. |
You may like...
System Assurances - Modeling and…
Prashant Johri, Adarsh Anand, …
Paperback
R2,610
Discovery Miles 26 100
Spatial Regression Analysis Using…
Daniel A. Griffith, Yongwan Chun, …
Paperback
R3,015
Discovery Miles 30 150
SAS Text Analytics for Business…
Teresa Jade, Biljana Belamaric-Wilsey, …
Hardcover
R2,569
Discovery Miles 25 690
Essential Java for Scientists and…
Brian Hahn, Katherine Malan
Paperback
R1,266
Discovery Miles 12 660
|