Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > Applications of computing > Databases > Data mining
This book reviews forecasting data mining models, from basic tools for stable data through causal models, to more advanced models using trends and cycles. These models are demonstrated on the basis of business-related data, including stock indices, crude oil prices, and the price of gold. The book's main approach is above all descriptive, seeking to explain how the methods concretely work; as such, it includes selected citations, but does not go into deep scholarly reference. The data sets and software reviewed were selected for their widespread availability to all readers with internet access.
This book constitutes the proceedings of the 24th International Symposium on Foundations of Intelligent Systems, ISMIS 2018, held in Limassol, Cyprus, in October 2018. The 32 full, 8 short, and 4 application papers presented in this volume were carefully reviewed and selected from 59 submissions. The papers deal with topics such as bioinformatics and health informatics, graph mining, image analysis, intelligent systems, mining complex patterns, novelty detection and class imbalance, social data analysis, spatio-temporal analysis, and topic modeling and opinion mining. In addition, three special sessions were organized, namely: Special Session on Granular and Soft Clustering for Data Science, Special Session on Intelligent Methodologies for Traffic Data Analysis and Mining, and Special Session on Advanced Methods in Machine Learning for Modeling Complex Data.
This book provides a practical approach to designing and implementing a Knowledge Management (KM) Strategy. The book explains how to design KM strategy so as to align business goals with KM objectives. The book also presents an approach for implementing KM strategy so as to make it sustainable. It covers all basic KM concepts, components of KM and the steps that are required for designing a KM strategy. As a result, the book can be used by beginners as well as practitioners. Knowledge management is a discipline that promotes an integrated approach to identifying, capturing, evaluating, retrieving, and sharing all of an enterprise's information assets. These assets may include databases, documents, policies, procedures, and previously un-captured expertise and experience in individual workers. Knowledge is considered to be the learning that results from experience and is embedded within individuals. Sometimes the knowledge is gained through critical thinking, watching others, and observing results of others. These observations then form a pattern which is converted in a 'generic form' to knowledge. This implies that knowledge can be formed only after data (which is generated through experience or observation) is grouped into information and then this information pattern is made generic wisdom. However, dissemination and acceptance of this knowledge becomes a key factor in knowledge management. The knowledge pyramid represents the usual concept of knowledge transformations, where data is transformed into information, and information is transformed into knowledge. Many organizations have struggled to manage knowledge and translate it into business benefits. This book is an attempt to show them how it can be done.
This book provides a comprehensive overview of the field of pattern mining with evolutionary algorithms. To do so, it covers formal definitions about patterns, patterns mining, type of patterns and the usefulness of patterns in the knowledge discovery process. As it is described within the book, the discovery process suffers from both high runtime and memory requirements, especially when high dimensional datasets are analyzed. To solve this issue, many pruning strategies have been developed. Nevertheless, with the growing interest in the storage of information, more and more datasets comprise such a dimensionality that the discovery of interesting patterns becomes a challenging process. In this regard, the use of evolutionary algorithms for mining pattern enables the computation capacity to be reduced, providing sufficiently good solutions. This book offers a survey on evolutionary computation with particular emphasis on genetic algorithms and genetic programming. Also included is an analysis of the set of quality measures most widely used in the field of pattern mining with evolutionary algorithms. This book serves as a review of the most important evolutionary algorithms for pattern mining. It considers the analysis of different algorithms for mining different type of patterns and relationships between patterns, such as frequent patterns, infrequent patterns, patterns defined in a continuous domain, or even positive and negative patterns. A completely new problem in the pattern mining field, mining of exceptional relationships between patterns, is discussed. In this problem the goal is to identify patterns which distribution is exceptionally different from the distribution in the complete set of data records. Finally, the book deals with the subgroup discovery task, a method to identify a subgroup of interesting patterns that is related to a dependent variable or target attribute. This subgroup of patterns satisfies two essential conditions: interpretability and interestingness.
Knowledge management (KM) is about managing the lifecycle of knowledge consisting of creating, storing, sharing and applying knowledge. Two main approaches towards KM are codification and personalization. The first focuses on capturing knowledge using technology and the latter on the process of socializing for sharing and creating knowledge. Social media are becoming very popular as individuals and also organizations learn how to use it. The primary applications of social media in a business context are marketing and recruitment. But there is also a huge potential for knowledge management in these organizations. For example, wikis can be used to collect organizational knowledge and social networking tools, which leads to exchanging new ideas and innovation. The interesting part of social media is that, by using them, one immediately starts to generate content that can be useful for the organization. Hence, they naturally combine the codification and personalisation approaches to KM. This book aims to provide an overview of new and innovative applications of social media and to report challenges that need to be solved. One example is the watering down of knowledge as a result of the use of organizational social media (Von Krogh, 2012).
This book investigates organizational learning from a variety of information processing perspectives. Continuous change and complexity in regulatory, social and economic environments are increasingly forcing organizations and their employees to acquire the necessary job-specific knowledge at the right time and in the right format. Though many regulatory documents are now available in digital form, their complexity and diversity make identifying the relevant elements for a particular context a challenging task. In such scenarios, business processes tend to be important sources of knowledge, containing rich but in many cases embedded, hidden knowledge. This book discusses the possible connection between business process models and corporate knowledge assets; knowledge extraction approaches based on organizational processes; developing and maintaining corporate knowledge bases; and semantic business process management and its relation to organizational learning approaches. The individual chapters reveal the different elements of a knowledge management solution designed to extract, organize and preserve the knowledge embedded in business processes so as to: enrich organizational knowledge bases in a systematic and controlled way, support employees in acquiring job role-specific knowledge, promote organizational learning, and steer human capital investment. All of these topics are analyzed on the basis of real-world cases from the domains of insurance, food safety, innovation, and funding.
This contributed volume explores the emerging intersection between big data analytics and genomics. Recent sequencing technologies have enabled high-throughput sequencing data generation for genomics resulting in several international projects which have led to massive genomic data accumulation at an unprecedented pace. To reveal novel genomic insights from this data within a reasonable time frame, traditional data analysis methods may not be sufficient or scalable, forcing the need for big data analytics to be developed for genomics. The computational methods addressed in the book are intended to tackle crucial biological questions using big data, and are appropriate for either newcomers or veterans in the field.This volume offers thirteen peer-reviewed contributions, written by international leading experts from different regions, representing Argentina, Brazil, China, France, Germany, Hong Kong, India, Japan, Spain, and the USA. In particular, the book surveys three main areas: statistical analytics, computational analytics, and cancer genome analytics. Sample topics covered include: statistical methods for integrative analysis of genomic data, computation methods for protein function prediction, and perspectives on machine learning techniques in big data mining of cancer. Self-contained and suitable for graduate students, this book is also designed for bioinformaticians, computational biologists, and researchers in communities ranging from genomics, big data, molecular genetics, data mining, biostatistics, biomedical science, cancer research, medical research, and biology to machine learning and computer science. Readers will find this volume to be an essential read for appreciating the role of big data in genomics, making this an invaluable resource for stimulating further research on the topic.
This successful book provides in its second edition an interactive and illustrative guide from two-dimensional curve fitting to multidimensional clustering and machine learning with neural networks or support vector machines. Along the way topics like mathematical optimization or evolutionary algorithms are touched. All concepts and ideas are outlined in a clear cut manner with graphically depicted plausibility arguments and a little elementary mathematics.The major topics are extensively outlined with exploratory examples and applications. The primary goal is to be as illustrative as possible without hiding problems and pitfalls but to address them. The character of an illustrative cookbook is complemented with specific sections that address more fundamental questions like the relation between machine learning and human intelligence.All topics are completely demonstrated with the computing platform Mathematica and the Computational Intelligence Packages (CIP), a high-level function library developed with Mathematica's programming language on top of Mathematica's algorithms. CIP is open-source and the detailed code used throughout the book is freely accessible.The target readerships are students of (computer) science and engineering as well as scientific practitioners in industry and academia who deserve an illustrative introduction. Readers with programming skills may easily port or customize the provided code. "'From curve fitting to machine learning' is ... a useful book. ... It contains the basic formulas of curve fitting and related subjects and throws in, what is missing in so many books, the code to reproduce the results.All in all this is an interesting and useful book both for novice as well as expert readers. For the novice it is a good introductory book and the expert will appreciate the many examples and working code". Leslie A. Piegl (Review of the first edition, 2012).
This book offers a comprehensive review of multilabel techniques widely used to classify and label texts, pictures, videos and music in the Internet. A deep review of the specialized literature on the field includes the available software needed to work with this kind of data. It provides the user with the software tools needed to deal with multilabel data, as well as step by step instruction on how to use them. The main topics covered are: * The special characteristics of multi-labeled data and the metrics available to measure them.* The importance of taking advantage of label correlations to improve the results.* The different approaches followed to face multi-label classification.* The preprocessing techniques applicable to multi-label datasets.* The available software tools to work with multi-label data. This book is beneficial for professionals and researchers in a variety of fields because of the wide range of potential applications for multilabel classification. Besides its multiple applications to classify different types of online information, it is also useful in many other areas, such as genomics and biology. No previous knowledge about the subject is required. The book introduces all the needed concepts to understand multilabel data characterization, treatment and evaluation.
This book introduces Meaningful Purposive Interaction Analysis (MPIA) theory, which combines social network analysis (SNA) with latent semantic analysis (LSA) to help create and analyse a meaningful learning landscape from the digital traces left by a learning community in the co-construction of knowledge. The hybrid algorithm is implemented in the statistical programming language and environment R, introducing packages which capture - through matrix algebra - elements of learners' work with more knowledgeable others and resourceful content artefacts. The book provides comprehensive package-by-package application examples, and code samples that guide the reader through the MPIA model to show how the MPIA landscape can be constructed and the learner's journey mapped and analysed. This building block application will allow the reader to progress to using and building analytics to guide students and support decision-making in learning.
This book reports on advanced theories and cutting-edge applications in the field of soft computing. The individual chapters, written by leading researchers, are based on contributions presented during the 4th World Conference on Soft Computing, held May 25-27, 2014, in Berkeley. The book covers a wealth of key topics in soft computing, focusing on both fundamental aspects and applications. The former include fuzzy mathematics, type-2 fuzzy sets, evolutionary-based optimization, aggregation and neural networks, while the latter include soft computing in data analysis, image processing, decision-making, classification, series prediction, economics, control, and modeling. By providing readers with a timely, authoritative view on the field, and by discussing thought-provoking developments and challenges, the book will foster new research directions in the diverse areas of soft computing.
This book summarizes the state-of-the-art in unsupervised learning. The contributors discuss how with the proliferation of massive amounts of unlabeled data, unsupervised learning algorithms, which can automatically discover interesting and useful patterns in such data, have gained popularity among researchers and practitioners. The authors outline how these algorithms have found numerous applications including pattern recognition, market basket analysis, web mining, social network analysis, information retrieval, recommender systems, market research, intrusion detection, and fraud detection. They present how the difficulty of developing theoretically sound approaches that are amenable to objective evaluation have resulted in the proposal of numerous unsupervised learning algorithms over the past half-century. The intended audience includes researchers and practitioners who are increasingly using unsupervised learning algorithms to analyze their data. Topics of interest include anomaly detection, clustering, feature extraction, and applications of unsupervised learning. Each chapter is contributed by a leading expert in the field.
This is the first textbook on attribute exploration, its theory, its algorithms forapplications, and some of its many possible generalizations. Attribute explorationis useful for acquiring structured knowledge through an interactive process, byasking queries to an expert. Generalizations that handle incomplete, faulty, orimprecise data are discussed, but the focus lies on knowledge extraction from areliable information source.The method is based on Formal Concept Analysis, a mathematical theory ofconcepts and concept hierarchies, and uses its expressive diagrams. The presentationis self-contained. It provides an introduction to Formal Concept Analysiswith emphasis on its ability to derive algebraic structures from qualitative data,which can be represented in meaningful and precise graphics.
This book introduces numerous algorithmic hybridizations between both worlds that show how machine learning can improve and support evolution strategies. The set of methods comprises covariance matrix estimation, meta-modeling of fitness and constraint functions, dimensionality reduction for search and visualization of high-dimensional optimization processes, and clustering-based niching. After giving an introduction to evolution strategies and machine learning, the book builds the bridge between both worlds with an algorithmic and experimental perspective. Experiments mostly employ a (1+1)-ES and are implemented in Python using the machine learning library scikit-learn. The examples are conducted on typical benchmark problems illustrating algorithmic concepts and their experimental behavior. The book closes with a discussion of related lines of research.
Numerical computation, knowledge discovery and statistical data analysis integrated with powerful 2D and 3D graphics for visualization are the key topics of this book. The Python code examples powered by the Java platform can easily be transformed to other programming languages, such as Java, Groovy, Ruby and BeanShell. This book equips the reader with a computational platform which, unlike other statistical programs, is not limited by a single programming language.The author focuses on practical programming aspects and covers a broad range of topics, from basic introduction to the Python language on the Java platform (Jython), to descriptive statistics, symbolic calculations, neural networks, non-linear regression analysis and many other data-mining topics. He discusses how to find regularities in real-world data, how to classify data, and how to process data for knowledge discoveries. The code snippets are so short that they easily fit into single pages. Numeric Computation and Statistical Data Analysis on the Java Platform is a great choice for those who want to learn how statistical data analysis can be done using popular programming languages, who want to integrate data analysis algorithms in full-scale applications, and deploy such calculations on the web pages or computational servers regardless of their operating system. It is an excellent reference for scientific computations to solve real-world problems using a comprehensive stack of open-source Java libraries included in the DataMelt (DMelt) project and will be appreciated by many data-analysis scientists, engineers and students.
This book constitutes the proceedings of the 15th International Workshop on Algorithms and Models for the Web Graph, WAW 2018, held in Moscow, Russia in May 2018. The 11 full papers presented in this volume were carefully reviewed and selected from various submissions. The papers focus on topics like the information retrieval and data mining on the Web; Web as a text repository and as a graph, induced in various ways by link among pages, hosts and users; the understanding of graphs that arise from the Web and various user activities on the Web; stimulation of the development of high-performance algorithms and applications that exploit these graphs.
This book starts with an introduction to process modeling and process paradigms, then explains how to query and analyze process models, and how to analyze the process execution data. In this way, readers receive a comprehensive overview of what is needed to identify, understand and improve business processes. The book chiefly focuses on concepts, techniques and methods. It covers a large body of knowledge on process analytics - including process data querying, analysis, matching and correlating process data and models - to help practitioners and researchers understand the underlying concepts, problems, methods, tools and techniques involved in modern process analytics. Following an introduction to basic business process and process analytics concepts, it describes the state of the art in this area before examining different analytics techniques in detail. In this regard, the book covers analytics over different levels of process abstractions, from process execution data and methods for linking and correlating process execution data, to inferring process models, querying process execution data and process models, and scalable process data analytics methods. In addition, it provides a review of commercial process analytics tools and their practical applications. The book is intended for a broad readership interested in business process management and process analytics. It provides researchers with an introduction to these fields by comprehensively classifying the current state of research, by describing in-depth techniques and methods, and by highlighting future research directions. Lecturers will find a wealth of material to choose from for a variety of courses, ranging from undergraduate courses in business process management to graduate courses in business process analytics. Lastly, it offers professionals a reference guide to the state of the art in commercial tools and techniques, complemented by many real-world use case scenarios.
This book offers an original and broad exploration of the fundamental methods in Clustering and Combinatorial Data Analysis, presenting new formulations and ideas within this very active field. With extensive introductions, formal and mathematical developments and real case studies, this book provides readers with a deeper understanding of the mutual relationships between these methods, which are clearly expressed with respect to three facets: logical, combinatorial and statistical. Using relational mathematical representation, all types of data structures can be handled in precise and unified ways which the author highlights in three stages: Clustering a set of descriptive attributes Clustering a set of objects or a set of object categories Establishing correspondence between these two dual clusterings Tools for interpreting the reasons of a given cluster or clustering are also included. Foundations and Methods in Combinatorial and Statistical Data Analysis and Clustering will be a valuable resource for students and researchers who are interested in the areas of Data Analysis, Clustering, Data Mining and Knowledge Discovery.
This book presents the proceedings of the 6th International Conference on Frontiers of Intelligent Computing: Theory and Applications (FICTA-2017), held in Bhubaneswar, Odisha. The event brought together researchers, scientists, engineers, and practitioners to exchange their new ideas and experiences in the domain of intelligent computing theories with prospective applications to various engineering disciplines. The book is divided into two volumes: Information and Decision Sciences, and Intelligent Engineering Informatics. This volume covers broad areas of Information and Decision Sciences, with papers exploring both the theoretical and practical aspects of data-intensive computing, data mining, evolutionary computation, knowledge management & networks, sensor networks, signal processing, wireless networks, protocols & architectures etc. The book also offers a valuable resource for students at the post-graduate level in various engineering disciplines.
Data clustering is a highly interdisciplinary field, the goal of which is to divide a set of objects into homogeneous groups such that objects in the same group are similar and objects in different groups are quite distinct. Thousands of theoretical papers and a number of books on data clustering have been published over the past 50 years. However, few books exist to teach people how to implement data clustering algorithms. This book was written for anyone who wants to implement or improve their data clustering algorithms. Using object-oriented design and programming techniques, Data Clustering in C++ exploits the commonalities of all data clustering algorithms to create a flexible set of reusable classes that simplifies the implementation of any data clustering algorithm. Readers can follow the development of the base data clustering classes and several popular data clustering algorithms. Additional topics such as data pre-processing, data visualization, cluster visualization, and cluster interpretation are briefly covered. This book is divided into three parts-- Data Clustering and C++ Preliminaries: A review of basic concepts of data clustering, the unified modeling language, object-oriented programming in C++, and design patterns A C++ Data Clustering Framework: The development of data clustering base classes Data Clustering Algorithms: The implementation of several popular data clustering algorithms A key to learning a clustering algorithm is to implement and experiment the clustering algorithm. Complete listings of classes, examples, unit test cases, and GNU configuration files are included in the appendices of this book as well as in the downloadable resources. The only requirements to compile the code are a modern C++ compiler and the Boost C++ libraries.
This book investigates the ways in which these systems can promote public value by encouraging the disclosure and reuse of privately-held data in ways that support collective values such as environmental sustainability. Supported by funding from the National Science Foundation, the authors' research team has been working on one such system, designed to enhance consumers ability to access information about the sustainability of the products that they buy and the supply chains that produce them. Pulled by rapidly developing technology and pushed by budget cuts, politicians and public managers are attempting to find ways to increase the public value of their actions. Policymakers are increasingly acknowledging the potential that lies in publicly disclosing more of the data that they hold, as well as incentivizing individuals and organizations to access, use, and combine it in new ways. Due to technological advances which include smarter phones, better ways to track objects and people as they travel, and more efficient data processing, it is now possible to build systems which use shared, transparent data in creative ways. The book adds to the current conversation among academics and practitioners about how to promote public value through data disclosure, focusing particularly on the roles that governments, businesses and non-profit actors can play in this process, making it of interest to both scholars and policy-makers.
This book describes analytical techniques for optimizing knowledge acquisition, processing, and propagation, especially in the contexts of cyber-infrastructure and big data. Further, it presents easy-to-use analytical models of knowledge-related processes and their applications. The need for such methods stems from the fact that, when we have to decide where to place sensors, or which algorithm to use for processing the data-we mostly rely on experts' opinions. As a result, the selected knowledge-related methods are often far from ideal. To make better selections, it is necessary to first create easy-to-use models of knowledge-related processes. This is especially important for big data, where traditional numerical methods are unsuitable. The book offers a valuable guide for everyone interested in big data applications: students looking for an overview of related analytical techniques, practitioners interested in applying optimization techniques, and researchers seeking to improve and expand on these techniques.
This book offers readers a comprehensive guide to the evolution of the database field from its earliest stages up to the present-and from classical relational database management systems to the current Big Data metaphor. In particular, it gathers the most significant research from the Italian database community that had relevant intersections with international projects. Big Data technology is currently dominating both the market and research. The book provides readers with a broad overview of key research efforts in modelling, querying and analysing data, which, over the last few decades, have became massive and heterogeneous areas.
The book gathers a collection of high-quality peer-reviewed research papers presented at the International Conference on Data and Information Systems (ICDIS 2017), held at Indira Gandhi National Tribal University, India from November 3 to 4, 2017. The book covers all aspects of computational sciences and information security. In chapters written by leading researchers, developers and practitioner from academia and industry, it highlights the latest developments and technical solutions, helping readers from the computer industry capitalize on key advances in next-generation computer and communication technology. |
You may like...
Wits University At 100 - From Excavation…
Wits Communications
Paperback
The Unresolved National Question - Left…
Edward Webster, Karin Pampallis
Paperback
(2)
Extremisms In Africa
Alain Tschudin, Stephen Buchanan-Clarke, …
Paperback
(1)
|