![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Databases > Data capture & analysis
This volume presents the latest advances in statistics and data science, including theoretical, methodological and computational developments and practical applications related to classification and clustering, data gathering, exploratory and multivariate data analysis, statistical modeling, and knowledge discovery and seeking. It includes contributions on analyzing and interpreting large, complex and aggregated datasets, and highlights numerous applications in economics, finance, computer science, political science and education. It gathers a selection of peer-reviewed contributions presented at the 16th Conference of the International Federation of Classification Societies (IFCS 2019), which was organized by the Greek Society of Data Analysis and held in Thessaloniki, Greece, on August 26-29, 2019.
This book presents a step by step Asset Health Management Optimization Approach Using Internet of Things (IoT). The authors provide a comprehensive study which includes the descriptive, diagnostic, predictive, and prescriptive analysis in detail. The presentation focuses on the challenges of the parameter selection, statistical data analysis, predictive algorithms, big data storage and selection, data pattern recognition, machine learning techniques, asset failure distribution estimation, reliability and availability enhancement, condition based maintenance policy, failure detection, data driven optimization algorithm, and a multi-objective optimization approach, all of which can significantly enhance the reliability and availability of the system.
This book constitutes the refereed joint proceedings of the 4th International Workshop on Large-Scale Annotation of Biomedical Data and Expert Label Synthesis, LABELS 2019, the First International Workshop on Hardware Aware Learning for Medical Imaging and Computer Assisted Intervention, HAL-MICCAI 2019, and the Second International Workshop on Correction of Brainshift with Intra-Operative Ultrasound, CuRIOUS 2019, held in conjunction with the 22nd International Conference on Medical Imaging and Computer-Assisted Intervention, MICCAI 2019, in Shenzhen, China, in October 2019. The 8 papers presented at LABELS 2019, the 5 papers presented at HAL-MICCAI 2019, and the 3 papers presented at CuRIOUS 2019 were carefully reviewed and selected from numerous submissions. The LABELS papers present a variety of approaches for dealing with a limited number of labels, from semi-supervised learning to crowdsourcing. The HAL-MICCAI papers cover a wide set of hardware applications in medical problems, including medical image segmentation, electron tomography, pneumonia detection, etc. The CuRIOUS papers provide a snapshot of the current progress in the field through extended discussions and provide researchers an opportunity to characterize their image registration methods on newly released standardized datasets of iUS-guided brain tumor resection.
Hybride Leistungsbundel (HLB) dienen dazu, ein innovatives und nutzenorientiertes Produktverstandnis von Sach- und Dienstleistungen zu etablieren. Hochkomplexe Anlagen lassen sich durch diese integrierte Betrachtung von Sach- und Dienstleistungsanteilen deutlich besser vermarkten. Der Band liefert einen Uberblick zu diesem Konzept und stellt entsprechende Methoden und Werkzeuge zur Entwicklung von Sach- und Dienstleistungen vor. Dabei berucksichtigen die Autoren den gesamten Zyklus: von der Planung und Entwicklung bis zur Erbringung und Nutzung."
This two-volume set, LNCS 11317 and 12318, constitutes the thoroughly refereed proceedings of the 4th International Joint Conference, APWeb-WAIM 2020, held in Tianjin, China, in September 2020. Due to the COVID-19 pandemic the conference was organizedas a fully online conference. The 42 full papers presented together with 17 short papers, and 6 demonstration papers were carefully reviewed and selected from 180 submissions. The papers are organized around the following topics: Big Data Analytics; Graph Data and Social Networks; Knowledge Graph; Recommender Systems; Information Extraction and Retrieval; Machine Learning; Blockchain; Data Mining; Text Analysis and Mining; Spatial, Temporal and Multimedia Databases; Database Systems; and Demo.
Images play a crucial role in shaping and reflecting political life. Digitization has vastly increased the presence of such images in daily life, creating valuable new research opportunities for social scientists. We show how recent innovations in computer vision methods can substantially lower the costs of using images as data. We introduce readers to the deep learning algorithms commonly used for object recognition, facial recognition, and visual sentiment analysis. We then provide guidance and specific instructions for scholars interested in using these methods in their own research.
Explore how User Research has been influenced over the years by a range of disciplines, such as HCI, usability, anthropology, cognitive psychology, ergonomics etc. This book aims to contribute to the User Research community and covers topics that will help UX professionals, students and stakeholders to gain a better understanding of what User Research is. Throughout the book you will acquire a practical skill set, ranging from how to get the research going, to building a case in order to receive the budget and resources needed. It will provide you with a clear account of how to organise your research, how to plan it, and how to manage stakeholders' expectations throughout the project. You'll see how to fit User Research into your organization and incorporate it through the different product development phases (Discovery, Alpha, Beta until Live), as well as how to grow a User Research team. Practical User Research reviews the methodologies used for User Research, looks at how to recruit participants along with how to collect and analyse data, finally focusing on how to interpret and present your findings. Cross-cultural research, accessibility and assisted digital research will also be discussed throughout this book. The final chapter gives you 10 project briefs, with which you will be able to apply your new skill set and put into practice what you have learnt. What You'll Learn Integrate user research into your business Apply user research to your product development cycle Review the appropriate processes necessary to carry out user research Take a pragmatic approach to user research, method by method Who This Book Is For Anyone that wants to understand more about user research.
Discover a variety of data-mining algorithms that are useful for selecting small sets of important features from among unwieldy masses of candidates, or extracting useful features from measured variables. As a serious data miner you will often be faced with thousands of candidate features for your prediction or classification application, with most of the features being of little or no value. You'll know that many of these features may be useful only in combination with certain other features while being practically worthless alone or in combination with most others. Some features may have enormous predictive power, but only within a small, specialized area of the feature space. The problems that plague modern data miners are endless. This book helps you solve this problem by presenting modern feature selection techniques and the code to implement them. Some of these techniques are: Forward selection component analysis Local feature selection Linking features and a target with a hidden Markov model Improvements on traditional stepwise selection Nominal-to-ordinal conversion All algorithms are intuitively justified and supported by the relevant equations and explanatory material. The author also presents and explains complete, highly commented source code. The example code is in C++ and CUDA C but Python or other code can be substituted; the algorithm is important, not the code that's used to write it. What You Will Learn Combine principal component analysis with forward and backward stepwise selection to identify a compact subset of a large collection of variables that captures the maximum possible variation within the entire set. Identify features that may have predictive power over only a small subset of the feature domain. Such features can be profitably used by modern predictive models but may be missed by other feature selection methods. Find an underlying hidden Markov model that controls the distributions of feature variables and the target simultaneously. The memory inherent in this method is especially valuable in high-noise applications such as prediction of financial markets. Improve traditional stepwise selection in three ways: examine a collection of 'best-so-far' feature sets; test candidate features for inclusion with cross validation to automatically and effectively limit model complexity; and at each step estimate the probability that our results so far could be just the product of random good luck. We also estimate the probability that the improvement obtained by adding a new variable could have been just good luck. Take a potentially valuable nominal variable (a category or class membership) that is unsuitable for input to a prediction model, and assign to each category a sensible numeric value that can be used as a model input. Who This Book Is For Intermediate to advanced data science programmers and analysts.
This book constitutes the proceedings of the 7th International Conference on Analysis of Images, Social Networks and Texts, AIST 2018, held in Moscow, Russia, in July 2018. The 29 full papers were carefully reviewed and selected from 107 submissions (of which 26 papers were rejected without being reviewed). The papers are organized in topical sections on natural language processing; analysis of images and video; general topics of data analysis; analysis of dynamic behavior through event data; optimization problems on graphs and network structures; and innovative systems.
Analyze vast amounts of data in record time using Apache Spark with Databricks in the Cloud. Learn the fundamentals, and more, of running analytics on large clusters in Azure and AWS, using Apache Spark with Databricks on top. Discover how to squeeze the most value out of your data at a mere fraction of what classical analytics solutions cost, while at the same time getting the results you need, incrementally faster. This book explains how the confluence of these pivotal technologies gives you enormous power, and cheaply, when it comes to huge datasets. You will begin by learning how cloud infrastructure makes it possible to scale your code to large amounts of processing units, without having to pay for the machinery in advance. From there you will learn how Apache Spark, an open source framework, can enable all those CPUs for data analytics use. Finally, you will see how services such as Databricks provide the power of Apache Spark, without you having to know anything about configuring hardware or software. By removing the need for expensive experts and hardware, your resources can instead be allocated to actually finding business value in the data. This book guides you through some advanced topics such as analytics in the cloud, data lakes, data ingestion, architecture, machine learning, and tools, including Apache Spark, Apache Hadoop, Apache Hive, Python, and SQL. Valuable exercises help reinforce what you have learned. What You Will Learn Discover the value of big data analytics that leverage the power of the cloud Get started with Databricks using SQL and Python in either Microsoft Azure or AWS Understand the underlying technology, and how the cloud and Apache Spark fit into the bigger picture See how these tools are used in the real world Run basic analytics, including machine learning, on billions of rows at a fraction of a cost or free Who This Book Is For Data engineers, data scientists, and cloud architects who want or need to run advanced analytics in the cloud. It is assumed that the reader has data experience, but perhaps minimal exposure to Apache Spark and Azure Databricks. The book is also recommended for people who want to get started in the analytics field, as it provides a strong foundation.
This book provides users with cutting edge methods and technologies in the area of big data and visual analytics, as well as an insight to the big data and data analytics research conducted by world-renowned researchers in this field. The authors present comprehensive educational resources on big data and visual analytics covering state-of-the art techniques on data analytics, data and information visualization, and visual analytics. Each chapter covers specific topics related to big data and data analytics as virtual data machine, security of big data, big data applications, high performance computing cluster, and big data implementation techniques. Every chapter includes a description of an unique contribution to the area of big data and visual analytics. This book is a valuable resource for researchers and professionals working in the area of big data, data analytics, and information visualization. Advanced-level students studying computer science will also find this book helpful as a secondary textbook or reference.
Cluster analysis finds groups in data automatically. Most methods have been heuristic and leave open such central questions as: how many clusters are there? Which method should I use? How should I handle outliers? Classification assigns new observations to groups given previously classified observations, and also has open questions about parameter tuning, robustness and uncertainty assessment. This book frames cluster analysis and classification in terms of statistical models, thus yielding principled estimation, testing and prediction methods, and sound answers to the central questions. It builds the basic ideas in an accessible but rigorous way, with extensive data examples and R code; describes modern approaches to high-dimensional data and networks; and explains such recent advances as Bayesian regularization, non-Gaussian model-based clustering, cluster merging, variable selection, semi-supervised and robust classification, clustering of functional data, text and images, and co-clustering. Written for advanced undergraduates in data science, as well as researchers and practitioners, it assumes basic knowledge of multivariate calculus, linear algebra, probability and statistics.
A four-color journey through a complete Tableau visualization Tableau is a popular data visualization tool that's easy for individual desktop use as well as enterprise. Used by financial analysts, marketers, statisticians, business and sales leadership, and many other job roles to present data visually for easy understanding, it's no surprise that Tableau is an essential tool in our data-driven economy. Visual Analytics with Tableau is a complete journey in Tableau visualization for a non-technical business user. You can start from zero, connect your first data, and get right into creating and publishing awesome visualizations and insightful dashboards. - Learn the different types of charts you can create - Use aggregation, calculated fields, and parameters - Create insightful maps - Share interactive dashboards Geared toward beginners looking to get their feet wet with Tableau, this book makes it easy and approachable to get started right away.
This handbook is a comprehensive reference guide for researchers, funding agencies and organizations engaged in survey research. Drawing on research from a world-class team of experts, this collection addresses the challenges facing survey-based data collection today as well as the potential opportunities presented by new approaches to survey research, including in the development of policy. It examines innovations in survey methodology and how survey scholars and practitioners should think about survey data in the context of the explosion of new digital sources of data. The Handbook is divided into four key sections: the challenges faced in conventional survey research; opportunities to expand data collection; methods of linking survey data with external sources; and, improving research transparency and data dissemination, with a focus on data curation, evaluating the usability of survey project websites, and the credibility of survey-based social science. Chapter 23 of this book is open access under a CC BY 4.0 license at link.springer.com.
The social sciences are becoming datafied. The questions once considered the domain of sociologists are now answered by data scientists operating on large datasets and breaking with methodological tradition, for better or worse. The traditional social sciences, such as sociology or anthropology, are under the double threat of becoming marginalized or even irrelevant, both from new methods of research which require more computational skills and from increasing competition from the corporate world which gains an additional advantage based on data access. However, unlike data scientists, sociologists and anthropologists have a long history of doing qualitative research. The more quantified datasets we have, the more difficult it is to interpret them without adding layers of qualitative interpretation. Big Data therefore needs Thick Data. This book presents the available arsenal of new methods and tools for studying society both quantitatively and qualitatively, opening ground for the social sciences to take the lead in analysing digital behaviour. It shows that Big Data can and should be supplemented and interpreted through thick data as well as cultural analysis. Thick Big Data is critically important for students and researchers in the social sciences to understand the possibilities of digital analysis, both in the quantitative and qualitative area, and to successfully build mixed-methods approaches.
Power BI Data Analysis and Visualization provides a roadmap to vendor choices and highlights why Microsoft's Power BI is a very viable, cost effective option for data visualization. The book covers the fundamentals and most commonly used features of Power BI, but also includes an in-depth discussion of advanced Power BI features such as natural language queries; embedding Power BI dashboards; and live streaming data. It discusses real solutions to extract data from the ERP application, Microsoft Dynamics CRM, and also offers ways to host the Power BI Dashboard as an Azure application, extracting data from popular data sources like Microsoft SQL Server and open-source PostgreSQL. Authored by Microsoft experts, this book uses real-world coding samples and screenshots to spotlight how to create reports, embed them in a webpage, view them across multiple platforms, and more. Business owners, IT professionals, data scientists, and analysts will benefit from this thorough presentation of Power BI and its functions.
Enterprise Resource Planning (ERP), Supply Chain Management (SCM), Customer Relationship Management (CRM), Business Intelligence (BI) and Big Data Analytics (BDA) are business related tasks and processes, which are supported by standardized software solutions. The book explains that this requires business oriented thinking and acting from IT specialists and data scientists. It is a good idea to let students experience this directly from the business perspective, for example as executives of a virtual company. The course simulates the stepwise integration of the linked business process chain ERP-SCM-CRM-BI-Big Data of four competing groups of companies. The course participants become board members with full P&L responsibility for business units of one of four beer brewery groups managing supply chains from production to retailer.
This book constitutes the proceedings of the 22nd Annual Conference on Research in Computational Molecular Biology, RECOMB 2018, held in Paris, France, in April 2018. The 16 extended and 22 short abstracts presented were carefully reviewed and selected from 193 submissions. The short abstracts are included in the back matter of the volume. They report on original research in all areas of computational molecular biology and bioinformatics.
Get started using Python in data analysis with this compact practical guide. This book includes three exercises and a case study on getting data in and out of Python code in the right format. Learn Data Analysis with Python also helps you discover meaning in the data using analysis and shows you how to visualize it. Each lesson is, as much as possible, self-contained to allow you to dip in and out of the examples as your needs dictate. If you are already using Python for data analysis, you will find a number of things that you wish you knew how to do in Python. You can then take these techniques and apply them directly to your own projects. If you aren't using Python for data analysis, this book takes you through the basics at the beginning to give you a solid foundation in the topic. As you work your way through the book you will have a better of idea of how to use Python for data analysis when you are finished. What You Will Learn Get data into and out of Python code Prepare the data and its format Find the meaning of the data Visualize the data using iPython Who This Book Is For Those who want to learn data analysis using Python. Some experience with Python is recommended but not required, as is some prior experience with data analysis or data science.
Ein Zeithistoriker und ein Fachwissenschaftler beschreiben die Entstehung der Wirtschaftsinformatik und ihre Entwicklung zur Wissenschaft: von den ersten wissenschaftlichen Analysen als Antwort auf die Probleme des Einsatzes der damals sogenannten EDV in den 1950er und 1960er Jahren bis hin zur Diskussion uber die Zukunft der Disziplin in einem globalen Wissenschaftssystem. 16Selbstzeugnisse von Wirtschaftsinformatikern der ersten Herausgeber-Generation der Zeitschrift Wirtschaftsinformatik werden von den Autoren dokumentiert und kommentiert."
This text introduces and provides instruction on the design and analysis of experiments for a broad audience. Formed by decades of teaching, consulting, and industrial experience in the Design of Experiments field, this new edition contains updated examples, exercises, and situations covering the science and engineering practice. This text minimizes the amount of mathematical detail, while still doing full justice to the mathematical rigor of the presentation and the precision of statements, making the text accessible for those who have little experience with design of experiments and who need some practical advice on using such designs to solve day-to-day problems. Additionally, an intuitive understanding of the principles is always emphasized, with helpful hints throughout.
Alpha-Versionen sind Lehrbucher, Gesetze, Hochglanzprospekte, Aktienneuemissionsanzeigen, Regierungserklarungen. Dahinter ist das Reale. Hinter den Lehrbuchern die vorlesende Forscherpersonlichkeit, hinter dem Prospekt der Rat des erfahrenen Fachverkaufers. Alpha-Versionen meiden Urteile, Meinungen und Leidenschaftlichkeit. Dieses Buch ist kompromisslos beta. Hier werden die schnellen Veranderungen der Informationsgesellschaft mit dem einhergehenden taglichen Wahnsinn aus moglichen und unmoglichen Perspektiven aufs Korn genommen - und wo es nicht anders geht, wird das zu arg Provozierende in Schwarzhumorsatire geniessbar gemacht ("Nicht nur zur Neujahrszeit" oder "Das Ende der DGeneration").Das Buch enthalt die bisherigen Texte der "Kult"-Kolumne Beta-inside (Informatik-Spektrum) des "Wild Duck" Autors, erganzt um Satiren, die eher "das Schonste" am Buche sind. Die Neuauflage wurde um ein Nachwort des Autors erweitert
Machine learning has finally come of age. With H2O software, you can perform machine learning and data analysis using a simple open source framework that's easy to use, has a wide range of OS and language support, and scales for big data. This hands-on guide teaches you how to use H20 with only minimal math and theory behind the learning algorithms. If you're familiar with R or Python, know a bit of statistics, and have some experience manipulating data, author Darren Cook will take you through H2O basics and help you conduct machine-learning experiments on different sample data sets. You'll explore several modern machine-learning techniques such as deep learning, random forests, unsupervised learning, and ensemble learning. Learn how to import, manipulate, and export data with H2O Explore key machine-learning concepts, such as cross-validation and validation data sets Work with three diverse data sets, including a regression, a multinomial classification, and a binomial classification Use H2O to analyze each sample data set with four supervised machine-learning algorithms Understand how cluster analysis and other unsupervised machine-learning algorithms work
This book constitutes revised selected papers from the 4th ECML PKDD Workshop on Data Analytics for Renewable Energy Integration, DARE 2016, held in Riva del Garda, Italy, in September 2016. The 11 papers presented in this volume were carefully reviewed and selected for inclusion in this book and handle topics such as time series forecasting, the detection of faults, cyber security, smart grid and smart cities, technology integration, demand response and many others.
This book constitutes revised selected papers from the third ECML PKDD Workshop on Data Analytics for Renewable Energy Integration, DARE 2015, held in Porto, Portugal, in September 2015. The 10 papers presented in this volume were carefully reviewed and selected for inclusion in this book. |
You may like...
Visual Communication - An Information…
Friedrich O. Huck, Carl L. Fales, …
Hardcover
R4,128
Discovery Miles 41 280
Enterprise Big Data Engineering…
Martin Atzmueller, Samia Oussena, …
Hardcover
R5,155
Discovery Miles 51 550
The Host in the Machine - Examining the…
Angela Thomas-Jones
Paperback
R1,318
Discovery Miles 13 180
Principles of 3D Image Analysis and…
Bernd Girod, Gunther Greiner, …
Hardcover
R4,271
Discovery Miles 42 710
|