![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer software packages > Other software packages > Mathematical & statistical software
Statistical Programming in SAS Second Edition provides a foundation for programming to implement statistical solutions using SAS, a system that has been used to solve data analytic problems for more than 40 years. The author includes motivating examples to inspire readers to generate programming solutions. Upper-level undergraduates, beginning graduate students, and professionals involved in generating programming solutions for data-analytic problems will benefit from this book. The ideal background for a reader is some background in regression modeling and introductory experience with computer programming. The coverage of statistical programming in the second edition includes Getting data into the SAS system, engineering new features, and formatting variables Writing readable and well-documented code Structuring, implementing, and debugging programs that are well documented Creating solutions to novel problems Combining data sources, extracting parts of data sets, and reshaping data sets as needed for other analyses Generating general solutions using macros Customizing output Producing insight-inspiring data visualizations Parsing, processing, and analyzing text Programming solutions using matrices and connecting to R Processing text Programming with matrices Connecting SAS with R Covering topics that are part of both base and certification exams.
* Targests readers with a background in programming, interested in an introduction/refresher in statistical hypothesis testing * Uses Python throughout * Provides the reader with the opportunity of using the book whenever needed rather than following a sequential path.
The objective of Kai Zhang and his research is to assess the existing process monitoring and fault detection (PM-FD) methods. His aim is to provide suggestions and guidance for choosing appropriate PM-FD methods, because the performance assessment study for PM-FD methods has become an area of interest in both academics and industry. The author first compares basic FD statistics, and then assesses different PM-FD methods to monitor the key performance indicators of static processes, steady-state dynamic processes and general dynamic processes including transient states. He validates the theoretical developments using both benchmark and real industrial processes.
This book provides new insights on the study of global environmental changes using the ecoinformatics tools and the adaptive-evolutionary technology of geoinformation monitoring. The main advantage of this book is that it gathers and presents extensive interdisciplinary expertise in the parameterization of global biogeochemical cycles and other environmental processes in the context of globalization and sustainable development. In this regard, the crucial global problems concerning the dynamics of the nature-society system are considered and the key problems of ensuring the system's sustainable development are studied. A new approach to the numerical modeling of the nature-society system is proposed and results are provided on modeling the dynamics of the system's characteristics with regard to scenarios of anthropogenic impacts on biogeochemical cycles, land ecosystems and oceans. The main purpose of this book is to develop a universal guide to information-modeling technologies for assessing the function of environmental subsystems under various climatic and anthropogenic conditions.
Clinical Data Quality Checks for CDISC Compliance using SAS is the first book focused on identifying and correcting data quality and CDISC compliance issues with real-world innovative SAS programming techniques such as Proc SQL, metadata and macro programming. Learn to master Proc SQL's subqueries and summary functions for multi-tasking process. Drawing on his more than 25 years' experience in the pharmaceutical industry, the author provides a unique approach that empowers SAS programmers to take control of data quality and CDISC compliance. This book helps you create a system of SDTM and ADaM checks that can be tracked for continuous improvement. How often have you encountered issues such as missing required variables, duplicate records, invalid derived variables and invalid sequence of two dates? With the SAS programming techniques introduced in this book, you can start to monitor these and more complex data and CDISC compliance issues. With increased standardization in SDTM and ADaM specifications and data values, codelist dictionaries can be created for better organization, planning and maintenance. This book includes a SAS program to create excel files containing unique values from all SDTM and ADaM variables as columns. In addition, another SAS program compares SDTM and ADaM codelist dictionaries with codelists from define.xml specifications. Having tools to automate this process greatly saves time from doing it manually. Features SDTMs and ADaMs Vitals SDTMs and ADaMs Data CDISC Specifications Compliance CDISC Data Compliance Protocol Compliance Codelist Dictionary Compliance
This book introduces advanced undergraduate, graduate students and practitioners to statistical methods for ranking data. An important aspect of nonparametric statistics is oriented towards the use of ranking data. Rank correlation is defined through the notion of distance functions and the notion of compatibility is introduced to deal with incomplete data. Ranking data are also modeled using a variety of modern tools such as CART, MCMC, EM algorithm and factor analysis. This book deals with statistical methods used for analyzing such data and provides a novel and unifying approach for hypotheses testing. The techniques described in the book are illustrated with examples and the statistical software is provided on the authors' website.
This textbook presents the basic concepts and methods of fluid mechanics, including Lagrangian and Eulerian descriptions, tensors of stresses and strains, continuity, momentum, energy, thermodynamics laws, and similarity theory. The models and their solutions are presented within a context of the mechanics of multiphase media. The treatment fully utilizes the computer algebra and software system Mathematica (R) to both develop concepts and help the reader to master modern methods of solving problems in fluid mechanics. Topics and features: Glossary of over thirty Mathematica (R) computer programs Extensive, self-contained appendix of Mathematica (R) functions and their use Chapter coverage of mechanics of multiphase heterogeneous media Detailed coverage of theory of shock waves in gas dynamics Thorough discussion of aerohydrodynamics of ideal and viscous fluids an d gases Complete worked examples with detailed solutions Problem-solving approach Foundations of Fluid Mechanics with Applications is a complete and accessible text or reference for graduates and professionals in mechanics, applied mathematics, physical sciences, materials science, and engineering. It is an essential resource for the study and use of modern solution methods for problems in fluid mechanics and the underlying mathematical models. The present, softcover reprint is designed to make this classic textbook available to a wider audience.
Describing novel mathematical concepts for recommendation engines, Realtime Data Mining: Self-Learning Techniques for Recommendation Engines features a sound mathematical framework unifying approaches based on control and learning theories, tensor factorization, and hierarchical methods. Furthermore, it presents promising results of numerous experiments on real-world data. The area of realtime data mining is currently developing at an exceptionally dynamic pace, and realtime data mining systems are the counterpart of today's "classic" data mining systems. Whereas the latter learn from historical data and then use it to deduce necessary actions, realtime analytics systems learn and act continuously and autonomously. In the vanguard of these new analytics systems are recommendation engines. They are principally found on the Internet, where all information is available in realtime and an immediate feedback is guaranteed. This monograph appeals to computer scientists and specialists in machine learning, especially from the area of recommender systems, because it conveys a new way of realtime thinking by considering recommendation tasks as control-theoretic problems. Realtime Data Mining: Self-Learning Techniques for Recommendation Engines will also interest application-oriented mathematicians because it consistently combines some of the most promising mathematical areas, namely control theory, multilevel approximation, and tensor factorization.
This volume compiles the major results of conference participants from the "Third International Conference in Network Analysis" held at the Higher School of Economics, Nizhny Novgorod in May 2013, with the aim to initiate further joint research among different groups. The contributions in this book cover a broad range of topics relevant to the theory and practice of network analysis, including the reliability of complex networks, software, theory, methodology, and applications. Network analysis has become a major research topic over the last several years. The broad range of applications that can be described and analyzed by means of a network has brought together researchers, practitioners from numerous fields such as operations research, computer science, transportation, energy, biomedicine, computational neuroscience and social sciences. In addition, new approaches and computer environments such as parallel computing, grid computing, cloud computing, and quantum computing have helped to solve large scale network optimization problems.
The book opens with a short introduction to Indian music, in particular classical Hindustani music, followed by a chapter on the role of statistics in computational musicology. The authors then show how to analyze musical structure using Rubato, the music software package for statistical analysis, in particular addressing modeling, melodic similarity and lengths, and entropy analysis; they then show how to analyze musical performance. Finally, they explain how the concept of seminatural composition can help a music composer to obtain the opening line of a raga-based song using Monte Carlo simulation. The book will be of interest to musicians and musicologists, particularly those engaged with Indian music.
Economists can use computer algebra systems to manipulate symbolic models, derive numerical computations, and analyze empirical relationships among variables. Maxima is an open-source multi-platform computer algebra system that rivals proprietary software. Maxima's symbolic and computational capabilities enable economists and financial analysts to develop a deeper understanding of models by allowing them to explore the implications of differences in parameter values, providing numerical solutions to problems that would be otherwise intractable, and by providing graphical representations that can guide analysis. This book provides a step-by-step tutorial for using this program to examine the economic relationships that form the core of microeconomics in a way that complements traditional modeling techniques. Readers learn how to phrase the relevant analysis and how symbolic expressions, numerical computations, and graphical representations can be used to learn from microeconomic models. In particular, comparative statics analysis is facilitated. Little has been published on Maxima and its applications in economics and finance, and this volume will appeal to advanced undergraduates, graduate-level students studying microeconomics, academic researchers in economics and finance, economists, and financial analysts.
This book highlights recent advances in natural computing, including biology and its theory, bio-inspired computing, computational aesthetics, computational models and theories, computing with natural media, philosophy of natural computing and educational technology. It presents extended versions of the best papers selected from the symposium "7th International Workshop on Natural Computing" (IWNC7), held in Tokyo, Japan, in 2013. The target audience is not limited to researchers working in natural computing but also those active in biological engineering, fine/media art design, aesthetics and philosophy.
Praise for the First Edition:
This book is a timely and critical introduction for those interested in what data science is (and isn't), and how it should be applied. The language is conversational and the content is accessible for readers without a quantitative or computational background; but, at the same time, it is also a practical overview of the field for the more technical readers. The overarching goal is to demystify the field and teach the reader how to develop an analytical mindset instead of following recipes. The book takes the scientist's approach of focusing on asking the right question at every step as this is the single most important factor contributing to the success of a data science project. Upon finishing this book, the reader should be asking more questions than I have answered. This book is, therefore, a practising scientist's approach to explaining data science through questions and examples.
This book gathers a selection of invited and contributed lectures from the European Conference on Numerical Mathematics and Advanced Applications (ENUMATH) held in Lausanne, Switzerland, August 26-30, 2013. It provides an overview of recent developments in numerical analysis, computational mathematics and applications from leading experts in the field. New results on finite element methods, multiscale methods, numerical linear algebra and discretization techniques for fluid mechanics and optics are presented. As such, the book offers a valuable resource for a wide range of readers looking for a state-of-the-art overview of advanced techniques, algorithms and results in numerical mathematics and scientific computing.
This book offers a snapshot of the state-of-the-art in classification at the interface between statistics, computer science and application fields. The contributions span a broad spectrum, from theoretical developments to practical applications; they all share a strong computational component. The topics addressed are from the following fields: Statistics and Data Analysis; Machine Learning and Knowledge Discovery; Data Analysis in Marketing; Data Analysis in Finance and Economics; Data Analysis in Medicine and the Life Sciences; Data Analysis in the Social, Behavioural, and Health Care Sciences; Data Analysis in Interdisciplinary Domains; Classification and Subject Indexing in Library and Information Science. The book presents selected papers from the Second European Conference on Data Analysis, held at Jacobs University Bremen in July 2014. This conference unites diverse researchers in the pursuit of a common topic, creating truly unique synergies in the process.
This edited volume on the latest advances in data science covers a wide range of topics in the context of data analysis and classification. In particular, it includes contributions on classification methods for high-dimensional data, clustering methods, multivariate statistical methods, and various applications. The book gathers a selection of peer-reviewed contributions presented at the Fifteenth Conference of the International Federation of Classification Societies (IFCS2015), which was hosted by the Alma Mater Studiorum, University of Bologna, from July 5 to 8, 2015.
The main focus of this book is on presenting advances in fuzzy statistics, and on proposing a methodology for testing hypotheses in the fuzzy environment based on the estimation of fuzzy confidence intervals, a context in which not only the data but also the hypotheses are considered to be fuzzy. The proposed method for estimating these intervals is based on the likelihood method and employs the bootstrap technique. A new metric generalizing the signed distance measure is also developed. In turn, the book presents two conceptually diverse applications in which defended intervals play a role: one is a novel methodology for evaluating linguistic questionnaires developed at the global and individual levels; the other is an extension of the multi-ways analysis of variance to the space of fuzzy sets. To illustrate these approaches, the book presents several empirical and simulation-based studies with synthetic and real data sets. In closing, it presents a coherent R package called "FuzzySTs" which covers all the previously mentioned concepts with full documentation and selected use cases. Given its scope, the book will be of interest to all researchers whose work involves advanced fuzzy statistical methods.
Contingency tables arise in diverse fields, including life sciences, education, social and political sciences, notably market research and opinion surveys. Their analysis plays an essential role in gaining insight into structures of the quantities under consideration and in supporting decision making. Combining both theory and applications, this book presents models and methods for the analysis of two- and multidimensional-contingency tables. An excellent reference for advanced undergraduates, graduate students, and practitioners in statistics as well as biosciences, social sciences, education, and economics, the work may also be used as a textbook for a course on categorical data analysis. Prerequisites include basic background on statistical inference and knowledge of statistical software packages.
The R Companion to Elementary Applied Statistics includes traditional applications covered in elementary statistics courses as well as some additional methods that address questions that might arise during or after the application of commonly used methods. Beginning with basic tasks and computations with R, readers are then guided through ways to bring data into R, manipulate the data as needed, perform common statistical computations and elementary exploratory data analysis tasks, prepare customized graphics, and take advantage of R for a wide range of methods that find use in many elementary applications of statistics. Features: Requires no familiarity with R or programming to begin using this book. Can be used as a resource for a project-based elementary applied statistics course, or for researchers and professionals who wish to delve more deeply into R. Contains an extensive array of examples that illustrate ideas on various ways to use pre-packaged routines, as well as on developing individualized code. Presents quite a few methods that may be considered non-traditional, or advanced. Includes accompanying carefully documented script files that contain code for all examples presented, and more. R is a powerful and free product that is gaining popularity across the scientific community in both the professional and academic arenas. Statistical methods discussed in this book are used to introduce the fundamentals of using R functions and provide ideas for developing further skills in writing R code. These ideas are illustrated through an extensive collection of examples. About the Author: Christopher Hay-Jahans received his Doctor of Arts in mathematics from Idaho State University in 1999. After spending three years at University of South Dakota, he moved to Juneau, Alaska, in 2002 where he has taught a wide range of undergraduate courses at University of Alaska Southeast.
This book constitutes the refereed proceedings of the 19th International Conference on Distributed and Computer and Communication Networks, DCCN 2016, held in Moscow, Russia, in November 2016. The 50 revised full papers and the 6 revised short papers presented were carefully reviewed and selected from 141 submissions. The papers cover the following topics: computer and communication networks architecture optimization; control in computer and communication networks; performance and QoS/QoE evaluation in wireless networks; analytical modeling and simulation of next-generation communications systems; queuing theory and reliability theory applications in computer networks; wireless 4G/5G networks, cm- and mm-wave radio technologies; RFID technology and its application in intellectual transportation networks; internet of things, wearables, and applications of distributed information systems; probabilistic and statistical models in information systems; mathematical modeling of high-tech systems; mathematical modeling and control problems; distributed and cloud computing systems, big data analytics.
The purpose of this handbook is to allow users to learn and master the mathematics software package MATLAB (R), as well as to serve as a quick reference to some of the most used instructions in the package. A unique feature of this handbook is that it can be used by the novice and by experienced users alike. For experienced users, it has four chapters with examples and applications in engineering, finance, physics, and optimization. Exercises are included, along with solutions available for the interested reader on the book's web page. These exercises are a complement for the interested reader who wishes to get a deeper understanding of MATLAB. Features Covers both MATLAB and introduction to Simulink Covers the use of GUIs in MATLAB and Simulink Offers downloadable examples and programs from the handbook's website Provides an introduction to object oriented programming using MATLAB Includes applications from many areas Includes the realization of executable files for MATLAB programs and Simulink models
This book discusses the problem of model choice when the statistical models are separate, also called nonnested. Chapter 1 provides an introduction, motivating examples and a general overview of the problem. Chapter 2 presents the classical or frequentist approach to the problem as well as several alternative procedures and their properties. Chapter 3 explores the Bayesian approach, the limitations of the classical Bayes factors and the proposed alternative Bayes factors to overcome these limitations. It also discusses a significance Bayesian procedure. Lastly, Chapter 4 examines the pure likelihood approach. Various real-data examples and computer simulations are provided throughout the text.
Learn how to develop powerful data analytics applications quickly for SQL Server database administrators and developers. Organizations will be able to sift data and derive the business intelligence needed to drive business decisions and profit. The addition of R to SQL Server 2016 places a powerful analytical processor into an environment most developers are already comfortable with - Visual Studio. This book walks even the newest of users through the creation process of a powerful R-language tool set for use in analyzing and reporting on your data. As a SQL Server database administrator or developer, it is sometimes difficult to stay on the bleeding edge of technology. Microsoft's addition of R to SQL Server 2016 is sure to be a game-changer, and the language will certainly become an integral part of future releases. R is in fact widely used today in statistical and related applications, and its use is only growing. Beginning SQL Server R Services helps you jump on board this important trend by providing good examples with detailed explanations of the WHY and not just the HOW. Walks you through setup and installation of SQL Server R Services. Explains the basics of working with R Tools for Visual Studio. Provides a road map to successfully creating custom R code. What You Will Learn Discover R's role in the SQL Server 2016 hierarchy. Manage the components needed to run SQL Server R Services code. Run R-language analytics and queries inside the database. Create analytic solutions that run across multiple datasets. Gain in-depth knowledge of the R language itself. Implement custom SQL Server R Services solutions. Who This Book Is For Any level of database administrator or developer, but specifically it's for those developers with the need to develop powerful data analytics applications quickly. Seasoned R developers will appreciate the book for its robust learning pattern, using visual aids in combination with properties explanations and scenarios. Beginning SQL Server R Services is the perfect "new hire" gift for new database developers in any organization.
Among the various multi-level formulations of mathematical models in decision making processes, this book focuses on the bi-level model. Being the most frequently used, the bi-level model addresses conflicts which exist in multi-level decision making processes. From the perspective of bi-level structure and uncertainty, this book takes real-life problems as the background, focuses on the so-called random-like uncertainty, and develops the general framework of random-like bi-level decision making problems. The random-like uncertainty considered in this book includes random phenomenon, random-overlapped random (Ra-Ra) phenomenon and fuzzy-overlapped random (Ra-Fu) phenomenon. Basic theory, models, algorithms and practical applications for different types of random-like bi-level decision making problems are also presented in this book. |
You may like...
I Shot Frank Zappa - My Life In…
Robert JH Davidson, John Elliott
Hardcover
R736
Discovery Miles 7 360
|