![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Business & Economics > Economics > Econometrics
This laboratory manual is intended for business analysts who wish to increase their skills in the use of statistical analysis to support business decisions. Most of the case studies use Excel,today's most common analysis tool. They range from the most basic descriptive analytical techniques to more advanced techniques such as linear regression and forecasting. Advanced projects cover inferential statistics for continuous variables (t-Test) and categorical variables (chi-square), as well as A/B testing. The manual ends with techniques to deal with the analysis of text data and tools to manage the analysis of large data sets (Big Data) using Excel. Includes companion files with solution spreadsheets, sample files, data sets, etc. from the book. Features: Teaches the statistical analysis skills needed to support business decisions Provides projects ranging from the most basic descriptive analytical techniques to more advanced techniques such as linear regression, forecasting, inferential statistics, and analyzing big data sets Includes companion files with solution spreadsheets, sample files, data sets, etc. used in the book's case studies
Volume 40 in the Advances in Econometrics series features twenty-three chapters that are split thematically into two parts. Part A presents novel contributions to the analysis of time series and panel data with applications in macroeconomics, finance, cognitive science and psychology, neuroscience, and labor economics. Part B examines innovations in stochastic frontier analysis, nonparametric and semiparametric modeling and estimation, A/B experiments, big-data analysis, and quantile regression. Individual chapters, written by both distinguished researchers and promising young scholars, cover many important topics in statistical and econometric theory and practice. Papers primarily, though not exclusively, adopt Bayesian methods for estimation and inference, although researchers of all persuasions should find considerable interest in the chapters contained in this work. The volume was prepared to honor the career and research contributions of Professor Dale J. Poirier. For researchers in econometrics, this volume includes the most up-to-date research across a wide range of topics.
Both theoretical and empirical aspects of single- and multi-winner voting procedures are presented in this collection of papers. Starting from a discussion of the underlying principles of democratic representation, the volume includes a description of a great variety of voting procedures. It lists and illustrates their susceptibility to the main voting paradoxes, assesses (under various models of voters' preferences) the probability of paradoxical outcomes, and discusses the relevance of the theoretical results to the choice of voting system.
Probability and Statistics have been widely used in various fields of science, including economics. Like advanced calculus and linear algebra, probability and statistics are indispensable mathematical tools in economics. Statistical inference in economics, namely econometric analysis, plays a crucial methodological role in modern economics, particularly in empirical studies in economics.This textbook covers probability theory and statistical theory in a coherent framework that will be useful in graduate studies in economics, statistics and related fields. As a most important feature, this textbook emphasizes intuition, explanations and applications of probability and statistics from an economic perspective.
Originally published in 1929. This balanced combination of fieldwork, statistical measurement, and realistic applications shows a synthesis of economics and political science in a conception of an organic relationship between the two sciences that involves functional analysis, institutional interpretation, and a more workmanlike approach to questions of organization such as division of labour and the control of industry. The treatise applies the test of fact through statistical analysis to economic and political theories for the quantitative and institutional approach in solving social and industrial problems. It constructs a framework of concepts, combining both economic and political theory, to systematically produce an original statement in general terms of the principles and methods for statistical fieldwork. The separation into Parts allows selective reading for the methods of statistical measurement; the principles and fallacies of applying these measures to economic and political fields; and the resultant construction of a statistical economics and politics. Basic statistical concepts are described for application, with each method of statistical measurement illustrated with instances relevant to the economic and political theory discussed and a statistical glossary is included.
Originally published in 1976 and with second edition published in 1984. This book established itself as the first genuinely introductory text on econometric methods, assuming no formal background on the part of the reader. The second edition maintains this distinctive feature. Fundamental concepts are carefully explained and, where possible, techniques are developed by verbal reasoning rather than formal proof. It provides all the material for a basic course. and is also ideal for a student working alone. Very little knowledge of maths and statistics is assumed, and the logic of statistical method is carefully stated. There are numerous exercises, designed to help the student assess individual progress. Methods are described with computer solutions in mind and the author shows how a variety of different calculations can be performed with relatively simple programs. This new edition also includes much new material - statistical tables are now included and their use carefully explained.
Originally published in 1984. Since the logic underlying economic theory can only be grasped fully by a thorough understanding of the mathematics, this book will be invaluable to economists wishing to understand vast areas of important research. It provides a basic introduction to the fundamental mathematical ideas of topology and calculus, and uses these to present modern singularity theory and recent results on the generic existence of isolated price equilibria in exchange economies.
Originally published in 1978. This book is designed to enable students on main courses in economics to comprehend literature which employs econometric techniques as a method of analysis, to use econometric techniques themselves to test hypotheses about economic relationships and to understand some of the difficulties involved in interpreting results. While the book is mainly aimed at second-year undergraduates undertaking courses in applied economics, its scope is sufficiently wide to take in students at postgraduate level who have no background in econometrics - it integrates fully the mathematical and statistical techniques used in econometrics with micro- and macroeconomic case studies.
Originally published in 1970; with a second edition in 1989. Empirical Bayes methods use some of the apparatus of the pure Bayes approach, but an actual prior distribution is assumed to generate the data sequence. It can be estimated thus producing empirical Bayes estimates or decision rules. In this second edition, details are provided of the derivation and the performance of empirical Bayes rules for a variety of special models. Attention is given to the problem of assessing the goodness of an empirical Bayes estimator for a given set of prior data. Chapters also focus on alternatives to the empirical Bayes approach and actual applications of empirical Bayes methods.
Originally published in 1985. Mathematical methods and models to facilitate the understanding of the processes of economic dynamics and prediction were refined considerably over the period before this book was written. The field had grown; and many of the techniques involved became extremely complicated. Areas of particular interest include optimal control, non-linear models, game-theoretic approaches, demand analysis and time-series forecasting. This book presents a critical appraisal of developments and identifies potentially productive new directions for research. It synthesises work from mathematics, statistics and economics and includes a thorough analysis of the relationship between system understanding and predictability.
Originally published in 1960 and 1966. This is an elementary introduction to the sources of economic statistics and their uses in answering economic questions. No mathematical knowledge is assumed, and no mathematical symbols are used. The book shows - by asking and answering a number of typical questions of applied economics - what the most useful statistics are, where they are found, and how they are to be interpreted and presented. The reader is introduced to the major British, European and American official sources, to the social accounts, to index numbers and averaging, and to elementary aids to inspection such as moving averages and scatter diagrams.
Herbert Scarf is a highly esteemed distinguished American economist. He is internationally famous for his early epoch-making work on optimal inventory policies and his highly influential study with Andrew Clark on optimal policies for a multi-echelon inventory problem, which initiated the important and flourishing field of supply chain management. Equally, he has gained world recognition for his classic study on the stability of the Walrasian price adjustment processes and his fundamental analysis on the relationship between the core and the set of competitive equilibria (the so-called Edgeworth conjecture). Further achievements include his remarkable sufficient condition for the existence of a core in non-transferable utility games and general exchange economies, his seminal paper with Lloyd Shapley on housing markets, and his pioneering study on increasing returns and models of production in the presence of indivisibilities. All in all, however, the name of Scarf is always remembered as a synonym for the computation of economic equilibria and fixed points. In the early 1960s he invented a path-breaking technique for computing equilibrium prices.This work has generated a major research field in economics termed Applied General Equilibrium Analysis and a corresponding area in operations research known as Simplicial Fixed Point Methods. This book comprises all his research articles and consists of four volumes. This volume collects Herbert Scarf's papers in the area of Operations Research and Management.
This book covers diverse themes, including institutions and efficiency, choice and values, law and economics, development and policy, and social and economic measurement. Written in honour of the distinguished economist Satish K. Jain, this compilation of essays should appeal not only to students and researchers of economic theory but also to those interested in the design and evaluation of institutions and policy.
Why should we be interested in macroeconomic survey expectations? This important book offers an in-depth treatment of this question from a point of view not covered in existing works on time-series econometrics and forecasting. Clements presents the nature of survey data, addresses some of the difficulties posed by the way in which survey expectations are elicited and considers the evaluation of point predictions and probability distributions. He outlines how, from a behavioural perspective, surveys offer insight into how economic agents form their expectations.
This volume deals with a range of contemporary issues in Indian and other world economies, with a focus on economic theory and policy and their longstanding implications. It analyses and predicts the mechanisms that can come into play to determine the function of institutions and the impact of public policy.
Financial econometrics is one of the greatest on-going success stories of recent decades, as it has become one of the most active areas of research in econometrics. In this book, Michael Clements presents a clear and logical explanation of the key concepts and ideas of forecasts of economic and financial variables. He shows that forecasts of the single most likely outcome of an economic and financial variable are of limited value. Forecasts that provide more information on the expected likely ranges of outcomes are more relevant. This book provides a comprehensive treatment of the evaluation of different types of forecasts and draws out the parallels between the different approaches. It describes the methods of evaluating these more complex forecasts which provide a fuller description of the range of possible future outcomes.
Numerical analysis is the study of computation and its accuracy, stability and often its implementation on a computer. This book focuses on the principles of numerical analysis and is intended to equip those readers who use statistics to craft their own software and to understand the advantages and disadvantages of different numerical methods.
The book provides an up-to-date survey of statistical and econometric techniques for the analysis of count data, with a focus on conditional distribution models. The book starts with a presentation of the benchmark Poisson regression model. Alternative models address unobserved heterogeneity, state dependence, selectivity, endogeneity, underreporting, and clustered sampling. Testing and estimation is discussed. Finally, applications are reviewed in various fields.
Spatial Econometrics provides a modern, powerful and flexible skillset to early career researchers interested in entering this rapidly expanding discipline. It articulates the principles and current practice of modern spatial econometrics and spatial statistics, combining rigorous depth of presentation with unusual depth of coverage. Introducing and formalizing the principles of, and 'need' for, models which define spatial interactions, the book provides a comprehensive framework for almost every major facet of modern science. Subjects covered at length include spatial regression models, weighting matrices, estimation procedures and the complications associated with their use. The work particularly focuses on models of uncertainty and estimation under various complications relating to model specifications, data problems, tests of hypotheses, along with systems and panel data extensions which are covered in exhaustive detail. Extensions discussing pre-test procedures and Bayesian methodologies are provided at length. Throughout, direct applications of spatial models are described in detail, with copious illustrative empirical examples demonstrating how readers might implement spatial analysis in research projects. Designed as a textbook and reference companion, every chapter concludes with a set of questions for formal or self--study. Finally, the book includes extensive supplementing information in a large sample theory in the R programming language that supports early career econometricians interested in the implementation of statistical procedures covered.
A Handbook of Statistical Analyses Using SPSS clearly describes how to conduct a range of univariate and multivariate statistical analyses using the latest version of the Statistical Package for the Social Sciences, SPSS 11. Each chapter addresses a different type of analytical procedure applied to one or more data sets, primarily from the social and behavioral sciences areas. Each chapter also contains exercises relating to the data sets introduced, providing readers with a means to develop both their SPSS and statistical skills. Model answers to the exercises are also provided. Readers can download all of the data sets from a companion Web site furnished by the authors.
This book focuses on the application of the partial hedging approach from modern math finance to equity-linked life insurance contracts. It provides an accessible, up-to-date introduction to quantifying financial and insurance risks. The book also explains how to price innovative financial and insurance products from partial hedging perspectives. Each chapter presents the problem, the mathematical formulation, theoretical results, derivation details, numerical illustrations, and references to further reading.
The interaction between mathematicians, statisticians and econometricians working in actuarial sciences and finance is producing numerous meaningful scientific results. This volume introduces new ideas, in the form of four-page papers, presented at the international conference Mathematical and Statistical Methods for Actuarial Sciences and Finance (MAF), held at Universidad Carlos III de Madrid (Spain), 4th-6th April 2018. The book covers a wide variety of subjects in actuarial science and financial fields, all discussed in the context of the cooperation between the three quantitative approaches. The topics include: actuarial models; analysis of high frequency financial data; behavioural finance; carbon and green finance; credit risk methods and models; dynamic optimization in finance; financial econometrics; forecasting of dynamical actuarial and financial phenomena; fund performance evaluation; insurance portfolio risk analysis; interest rate models; longevity risk; machine learning and soft-computing in finance; management in insurance business; models and methods for financial time series analysis, models for financial derivatives; multivariate techniques for financial markets analysis; optimization in insurance; pricing; probability in actuarial sciences, insurance and finance; real world finance; risk management; solvency analysis; sovereign risk; static and dynamic portfolio selection and management; trading systems. This book is a valuable resource for academics, PhD students, practitioners, professionals and researchers, and is also of interest to other readers with quantitative background knowledge.
This book explores new topics in modern research on empirical corporate finance and applied accounting, especially the econometric analysis of microdata. Dubbed "financial microeconometrics" by the author, this concept unites both methodological and applied approaches. The book examines how quantitative methods can be applied in corporate finance and accounting research in order to predict companies getting into financial distress. Presented in a clear and straightforward manner, it also suggests methods for linking corporate governance to financial performance, and discusses what the determinants of accounting disclosures are. Exploring these questions by way of numerous practical examples, this book is intended for researchers, practitioners and students who are not yet familiar with the variety of approaches available for data analysis and microeconometrics. "This book on financial microeconometrics is an excellent starting point for research in corporate finance and accounting. In my view, the text is positioned between a narrative and a scientific treatise. It is based on a vast amount of literature but is not overloaded with formulae. My appreciation of financial microeconometrics has very much increased. The book is well organized and properly written. I enjoyed reading it." Wolfgang Marty, Senior Investment Strategist, AgaNola AG
This book employs a computable general equilibrium (CGE) model - a widely used economic model which uses actual data to provide economic analysis and policy assessment - and applies it to economic data on Singapore's tourism industry. The authors set out to demonstrate how a novice modeller can acquire the necessary skills and knowledge to successfully apply general equilibrium models to tourism studies. The chapters explain how to build a computable general equilibrium model for tourism, how to conduct simulation and, most importantly, how to analyse modelling results. This applied study acts as a modelling book at both introductory and intermediate levels, specifically targeting students and researchers who are interested in and wish to learn computable general equilibrium modelling. The authors offer insightful analysis of Singapore's tourism industry and provide both students and researchers with a guide on how to apply general equilibrium models to actual economic data and draw accurate conclusions.
Herbert Scarf is a highly esteemed distinguished American economist. He is internationally famous for his early epoch-making work on optimal inventory policies and his highly influential study with Andrew Clark on optimal policies for a multi-echelon inventory problem, which initiated the important and flourishing field of supply chain management. Equally, he has gained world recognition for his classic study on the stability of the Walrasian price adjustment processes and his fundamental analysis on the relationship between the core and the set of competitive equilibria (the so-called Edgeworth conjecture). Further achievements include his remarkable sufficient condition for the existence of a core in non-transferable utility games and general exchange economies, his seminal paper with Lloyd Shapley on housing markets, and his pioneering study on increasing returns and models of production in the presence of indivisibilities. All in all, however, the name of Scarf is always remembered as a synonym for the computation of economic equilibria and fixed points. In the early 1960s he invented a path-breaking technique for computing equilibrium prices. This work has generated a major research field in economics termed Applied General Equilibrium Analysis and a corresponding area in operations research known as Simplicial Fixed Point Methods. This book comprises all his research articles and consists of four volumes. This volume collects Herbert Scarf's papers in the area of Economics and Game Theory. |
You may like...
The Oxford Handbook of the Economics of…
Yann Bramoulle, Andrea Galeotti, …
Hardcover
R5,455
Discovery Miles 54 550
Operations And Supply Chain Management
David Collier, James Evans
Hardcover
The Handbook of Historical Economics
Alberto Bisin, Giovanni Federico
Paperback
R2,567
Discovery Miles 25 670
Introductory Econometrics - A Modern…
Jeffrey Wooldridge
Hardcover
Linear and Non-Linear Financial…
Mehmet Kenan Terzioglu, Gordana Djurovic
Hardcover
R3,581
Discovery Miles 35 810
Operations and Supply Chain Management
James Evans, David Collier
Hardcover
|