![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Applications of computing > Databases > Data mining
This book discusses the challenges facing current research in knowledge discovery and data mining posed by the huge volumes of complex data now gathered in various real-world applications (e.g., business process monitoring, cybersecurity, medicine, language processing, and remote sensing). The book consists of 14 chapters covering the latest research by the authors and the research centers they represent. It illustrates techniques and algorithms that have recently been developed to preserve the richness of the data and allow us to efficiently and effectively identify the complex information it contains. Presenting the latest developments in complex pattern mining, this book is a valuable reference resource for data science researchers and professionals in academia and industry.
The present text aims at helping the reader to maximize the reuse of information. Topics covered include tools and services for creating simple, rich, and reusable knowledge representations to explore strategies for integrating this knowledge into legacy systems. The reuse and integration are essential concepts that must be enforced to avoid duplicating the effort and reinventing the wheel each time in the same field. This problem is investigated from different perspectives. in organizations, high volumes of data from different sources form a big threat for filtering out the information for effective decision making. the reader will be informed of the most recent advances in information reuse and integration.
This proceedings volume introduces recent work on the storage, retrieval and visualization of spatial Big Data, data-intensive geospatial computing and related data quality issues. Further, it addresses traditional topics such as multi-scale spatial data representations, knowledge discovery, space-time modeling, and geological applications. Spatial analysis and data mining are increasingly facing the challenges of Big Data as more and more types of crowd sourcing spatial data are used in GIScience, such as movement trajectories, cellular phone calls, and social networks. In order to effectively manage these massive data collections, new methods and algorithms are called for. The book highlights state-of-the-art advances in the handling and application of spatial data, especially spatial Big Data, offering a cutting-edge reference guide for graduate students, researchers and practitioners in the field of GIScience.
Text mining applications have experienced tremendous advances because of web 2.0 and social networking applications. Recent advances in hardware and software technology have lead to a number of unique scenarios where text mining algorithms are learned. Mining Text Data introduces an important niche in the text analytics field, and is an edited volume contributed by leading international researchers and practitioners focused on social networks & data mining. This book contains a wide swath in topics across social networks & data mining. Each chapter contains a comprehensive survey including the key research content on the topic, and the future directions of research in the field. There is a special focus on Text Embedded with Heterogeneous and Multimedia Data which makes the mining process much more challenging. A number of methods have been designed such as transfer learning and cross-lingual mining for such cases. Mining Text Data simplifies the content, so that advanced-level students, practitioners and researchers in computer science can benefit from this book. Academic and corporate libraries, as well as ACM, IEEE, and Management Science focused on information security, electronic commerce, databases, data mining, machine learning, and statistics are the primary buyers for this reference book.
This unique text/reference describes an exciting and novel approach to supercomputing in the DataFlow paradigm. The major advantages and applications of this approach are clearly described, and a detailed explanation of the programming model is provided using simple yet effective examples. The work is developed from a series of lecture courses taught by the authors in more than 40 universities across more than 20 countries, and from research carried out by Maxeler Technologies, Inc. Topics and features: presents a thorough introduction to DataFlow supercomputing for big data problems; reviews the latest research on the DataFlow architecture and its applications; introduces a new method for the rapid handling of real-world challenges involving large datasets; provides a case study on the use of the new approach to accelerate the Cooley-Tukey algorithm on a DataFlow machine; includes a step-by-step guide to the web-based integrated development environment WebIDE.
Data mining is a very active research area with many successful real-world app- cations. It consists of a set of concepts and methods used to extract interesting or useful knowledge (or patterns) from real-world datasets, providing valuable support for decision making in industry, business, government, and science. Although there are already many types of data mining algorithms available in the literature, it is still dif cult for users to choose the best possible data mining algorithm for their particular data mining problem. In addition, data mining al- rithms have been manually designed; therefore they incorporate human biases and preferences. This book proposes a new approach to the design of data mining algorithms. - stead of relying on the slow and ad hoc process of manual algorithm design, this book proposes systematically automating the design of data mining algorithms with an evolutionary computation approach. More precisely, we propose a genetic p- gramming system (a type of evolutionary computation method that evolves c- puter programs) to automate the design of rule induction algorithms, a type of cl- si cation method that discovers a set of classi cation rules from data. We focus on genetic programming in this book because it is the paradigmatic type of machine learning method for automating the generation of programs and because it has the advantage of performing a global search in the space of candidate solutions (data mining algorithms in our case), but in principle other types of search methods for this task could be investigated in the future.
This book is the first work that systematically describes the procedure of data mining and knowledge discovery on Bioinformatics databases by using the state-of-the-art hierarchical feature selection algorithms. The novelties of this book are three-fold. To begin with, this book discusses the hierarchical feature selection in depth, which is generally a novel research area in Data Mining/Machine Learning. Seven different state-of-the-art hierarchical feature selection algorithms are discussed and evaluated by working with four types of interpretable classification algorithms (i.e. three types of Bayesian network classification algorithms and the k-nearest neighbours classification algorithm). Moreover, this book discusses the application of those hierarchical feature selection algorithms on the well-known Gene Ontology database, where the entries (terms) are hierarchically structured. Gene Ontology database that unifies the representations of gene and gene products annotation provides the resource for mining valuable knowledge about certain biological research topics, such as the Biology of Ageing. Furthermore, this book discusses the mined biological patterns by the hierarchical feature selection algorithms relevant to the ageing-associated genes. Those patterns reveal the potential ageing-associated factors that inspire future research directions for the Biology of Ageing research.
This book is about inductive databases and constraint-based data mining, emerging research topics lying at the intersection of data mining and database research. The aim of the book as to provide an overview of the state-of- the art in this novel and - citing research area. Of special interest are the recent methods for constraint-based mining of global models for prediction and clustering, the uni?cation of pattern mining approaches through constraint programming, the clari?cation of the re- tionship between mining local patterns and global models, and the proposed in- grative frameworks and approaches for inducive databases. On the application side, applications to practically relevant problems from bioinformatics are presented. Inductive databases (IDBs) represent a database view on data mining and kno- edge discovery. IDBs contain not only data, but also generalizations (patterns and models) valid in the data. In an IDB, ordinary queries can be used to access and - nipulate data, while inductive queries can be used to generate (mine), manipulate, and apply patterns and models. In the IDB framework, patterns and models become "?rst-class citizens" and KDD becomes an extended querying process in which both the data and the patterns/models that hold in the data are queried.
Data Mining introduces in clear and simple ways how to use existing data mining methods to obtain effective solutions for a variety of management and engineering design problems. Data Mining is organised into two parts: the first provides a focused introduction to data mining and the second goes into greater depth on subjects such as customer analysis. It covers almost all managerial activities of a company, including: * supply chain design, * product development, * manufacturing system design, * product quality control, and * preservation of privacy. Incorporating recent developments of data mining that have made it possible to deal with management and engineering design problems with greater efficiency and efficacy, Data Mining presents a number of state-of-the-art topics. It will be an informative source of information for researchers, but will also be a useful reference work for industrial and managerial practitioners.
With the healthcare industry becoming increasingly more competitive, there exists a need for medical institutions to improve both the efficiency and the quality of their services. In order to do so, it is important to investigate how statistical models can be used to study health outcomes. Cases on Health Outcomes and Clinical Data Mining: Studies and Frameworks provides several case studies developed by faculty and graduates of the University of Louisville's PhD program in Applied and Industrial Mathematics. The studies in this book use non-traditional, exploratory data analysis and data mining tools to examine health outcomes, finding patterns and trends in observational data. This book is ideal for the next generation of data mining practitioners.
"Incomplete Information System and Rough Set Theory: Models and Attribute Reductions" covers theoretical study of generalizations of rough set model in various incomplete information systems. It discusses not only the regular attributes but also the criteria in the incomplete information systems. Based on different types of rough set models, the book presents the practical approaches to compute several reducts in terms of these models. The book is intended for researchers and postgraduate students in machine learning, data mining and knowledge discovery, especially for those who are working in rough set theory, and granular computing. Dr. Xibei Yang is a lecturer at the School of Computer Science and Engineering, Jiangsu University of Science and Technology, China; Jingyu Yang is a professor at the School of Computer Science, Nanjing University of Science and Technology, China.
Recent advances in high-throughput technologies have resulted in a deluge of biological information. Yet the storage, analysis, and interpretation of such multifaceted data require effective and efficient computational tools. This unique text/reference addresses the need for a unified framework describing how soft computing and machine learning techniques can be judiciously formulated and used in building efficient pattern recognition models. The book reviews both established and cutting-edge research, following a clear structure reflecting the major phases of a pattern recognition system: classification, feature selection, and clustering. The text provides a careful balance of theory, algorithms, and applications, with a particular emphasis given to applications in computational biology and bioinformatics. Topics and features: reviews the development of scalable pattern recognition algorithms for computational biology and bioinformatics; integrates different soft computing and machine learning methodologies with pattern recognition tasks; discusses in detail the integration of different techniques for handling uncertainties in decision-making and efficiently mining large biological datasets; presents a particular emphasis on real-life applications, such as microarray expression datasets and magnetic resonance images; includes numerous examples and experimental results to support the theoretical concepts described; concludes each chapter with directions for future research and a comprehensive bibliography. This important work will be of great use to graduate students and researchers in the fields of computer science, electrical and biomedical engineering. Researchers and practitioners involved in pattern recognition, machine learning, computational biology and bioinformatics, data mining, and soft computing will also find the book invaluable.
"Turn yourself into a Data Head. You'll become a more valuable employee and make your organization more successful." Thomas H. Davenport, Research Fellow, Author of Competing on Analytics, Big Data @ Work, and The AI Advantage You've heard the hype around data--now get the facts. In Becoming a Data Head: How to Think, Speak, and Understand Data Science, Statistics, and Machine Learning, award-winning data scientists Alex Gutman and Jordan Goldmeier pull back the curtain on data science and give you the language and tools necessary to talk and think critically about it. You'll learn how to: Think statistically and understand the role variation plays in your life and decision making Speak intelligently and ask the right questions about the statistics and results you encounter in the workplace Understand what's really going on with machine learning, text analytics, deep learning, and artificial intelligence Avoid common pitfalls when working with and interpreting data Becoming a Data Head is a complete guide for data science in the workplace: covering everything from the personalities you'll work with to the math behind the algorithms. The authors have spent years in data trenches and sought to create a fun, approachable, and eminently readable book. Anyone can become a Data Head--an active participant in data science, statistics, and machine learning. Whether you're a business professional, engineer, executive, or aspiring data scientist, this book is for you.
This volume covers some of the topics that are related to the rapidly growing field of biomedical informatics. In June 11-12, 2010 a workshop entitled 'Optimization and Data Analysis in Biomedical Informatics' was organized at The Fields Institute. Following this event invited contributions were gathered based on the talks presented at the workshop, and additional invited chapters were chosen from world's leading experts. In this publication, the authors share their expertise in the form of state-of-the-art research and review chapters, bringing together researchers from different disciplines and emphasizing the value of mathematical methods in the areas of clinical sciences. This work is targeted to applied mathematicians, computer scientists, industrial engineers, and clinical scientists who are interested in exploring emerging and fascinating interdisciplinary topics of research. It is designed to further stimulate and enhance fruitful collaborations between scientists from different disciplines.
This edited volume gathers the proceedings of the Symposium GIS Ostrava 2016, the Rise of Big Spatial Data, held at the Technical University of Ostrava, Czech Republic, March 16-18, 2016. Combining theoretical papers and applications by authors from around the globe, it summarises the latest research findings in the area of big spatial data and key problems related to its utilisation. Welcome to dawn of the big data era: though it's in sight, it isn't quite here yet. Big spatial data is characterised by three main features: volume beyond the limit of usual geo-processing, velocity higher than that available using conventional processes, and variety, combining more diverse geodata sources than usual. The popular term denotes a situation in which one or more of these key properties reaches a point at which traditional methods for geodata collection, storage, processing, control, analysis, modelling, validation and visualisation fail to provide effective solutions. >Entering the era of big spatial data calls for finding solutions that address all "small data" issues that soon create "big data" troubles. Resilience for big spatial data means solving the heterogeneity of spatial data sources (in topics, purpose, completeness, guarantee, licensing, coverage etc.), large volumes (from gigabytes to terabytes and more), undue complexity of geo-applications and systems (i.e. combination of standalone applications with web services, mobile platforms and sensor networks), neglected automation of geodata preparation (i.e. harmonisation, fusion), insufficient control of geodata collection and distribution processes (i.e. scarcity and poor quality of metadata and metadata systems), limited analytical tool capacity (i.e. domination of traditional causal-driven analysis), low visual system performance, inefficient knowledge-discovery techniques (for transformation of vast amounts of information into tiny and essential outputs) and much more. These trends are accelerating as sensors become more ubiquitous around the world.
On various examples ranging from geosciences to environmental sciences, this book explains how to generate an adequate description of uncertainty, how to justify semiheuristic algorithms for processing uncertainty, and how to make these algorithms more computationally efficient. It explains in what sense the existing approach to uncertainty as a combination of random and systematic components is only an approximation, presents a more adequate three-component model with an additional periodic error component, and explains how uncertainty propagation techniques can be extended to this model. The book provides a justification for a practically efficient heuristic technique (based on fuzzy decision-making). It explains how the computational complexity of uncertainty processing can be reduced. The book also shows how to take into account that in real life, the information about uncertainty is often only partially known, and, on several practical examples, explains how to extract the missing information about uncertainty from the available data.
This book presents advances in matrix and tensor data processing in
the domain of signal, image and information processing. The
theoretical mathematical approaches are discusses in the context of
potential applications in sensor and cognitive systems engineering.
Data mining has emerged as one of the most active areas in information and c- munication technologies(ICT). With the boomingof the global economy, and ub- uitouscomputingandnetworkingacrosseverysectorand business, data andits deep analysis becomes a particularly important issue for enhancing the soft power of an organization, its production systems, decision-making and performance. The last ten years have seen ever-increasingapplications of data mining in business, gove- ment, social networks and the like. However, a crucial problem that prevents data mining from playing a strategic decision-support role in ICT is its usually limited decision-support power in the real world. Typical concerns include its actionability, workability, transferability, and the trustworthy, dependable, repeatable, operable and explainable capabilities of data mining algorithms, tools and outputs. This monograph, Domain Driven Data Mining, is motivated by the real-world challenges to and complexities of the current KDD methodologies and techniques, which are critical issues faced by data mining, as well as the ?ndings, thoughts and lessons learned in conducting several large-scale real-world data mining bu- ness applications. The aim and objective of domain driven data mining is to study effective and ef?cient methodologies, techniques, tools, and applications that can discover and deliver actionable knowledge that can be passed on to business people for direct decision-making and action-takin
This book presents Explainable Artificial Intelligence (XAI), which aims at producing explainable models that enable human users to understand and appropriately trust the obtained results. The authors discuss the challenges involved in making machine learning-based AI explainable. Firstly, that the explanations must be adapted to different stakeholders (end-users, policy makers, industries, utilities etc.) with different levels of technical knowledge (managers, engineers, technicians, etc.) in different application domains. Secondly, that it is important to develop an evaluation framework and standards in order to measure the effectiveness of the provided explanations at the human and the technical levels. This book gathers research contributions aiming at the development and/or the use of XAI techniques in order to address the aforementioned challenges in different applications such as healthcare, finance, cybersecurity, and document summarization. It allows highlighting the benefits and requirements of using explainable models in different application domains in order to provide guidance to readers to select the most adapted models to their specified problem and conditions. Includes recent developments of the use of Explainable Artificial Intelligence (XAI) in order to address the challenges of digital transition and cyber-physical systems; Provides a textual scientific description of the use of XAI in order to address the challenges of digital transition and cyber-physical systems; Presents examples and case studies in order to increase transparency and understanding of the methodological concepts.
The book proposes new technologies and discusses future solutions for design infrastructure for ICT. The book contains high quality submissions presented at Second International Conference on Information and Communication Technology for Sustainable Development (ICT4SD - 2016) held at Goa, India during 1 - 2 July, 2016. The conference stimulates the cutting-edge research discussions among many academic pioneering researchers, scientists, industrial engineers, and students from all around the world. The topics covered in this book also focus on innovative issues at international level by bringing together the experts from different countries.
This book examines the techniques and applications involved in the Web Mining, Web Personalization and Recommendation and Web Community Analysis domains, including a detailed presentation of the principles, developed algorithms, and systems of the research in these areas. The applications of web mining, and the issue of how to incorporate web mining into web personalization and recommendation systems are also reviewed. Additionally, the volume explores web community mining and analysis to find the structural, organizational and temporal developments of web communities and reveal the societal sense of individuals or communities. The volume will benefit both academic and industry communities interested in the techniques and applications of web search, web data management, web mining and web knowledge discovery, as well as web community and social network analysis.
The advances of live cell video imaging and high-throughput technologies for functional and chemical genomics provide unprecedented opportunities to understand how biological processes work in subcellularand multicellular systems. The interdisciplinary research field of Video Bioinformatics is defined by BirBhanu as the automated processing, analysis, understanding, data mining, visualization, query-basedretrieval/storage of biological spatiotemporal events/data and knowledge extracted from dynamic imagesand microscopic videos. Video bioinformatics attempts to provide a deeper understanding of continuousand dynamic life processes.Genome sequences alone lack spatial and temporal information, and video imaging of specific moleculesand their spatiotemporal interactions, using a range of imaging methods, are essential to understandhow genomes create cells, how cells constitute organisms, and how errant cells cause disease. The bookexamines interdisciplinary research issues and challenges with examples that deal with organismal dynamics,intercellular and tissue dynamics, intracellular dynamics, protein movement, cell signaling and softwareand databases for video bioinformatics.Topics and Features* Covers a set of biological problems, their significance, live-imaging experiments, theory andcomputational methods, quantifiable experimental results and discussion of results.* Provides automated methods for analyzing mild traumatic brain injury over time, identifying injurydynamics after neonatal hypoxia-ischemia and visualizing cortical tissue changes during seizureactivity as examples of organismal dynamics* Describes techniques for quantifying the dynamics of human embryonic stem cells with examplesof cell detection/segmentation, spreading and other dynamic behaviors which are important forcharacterizing stem cell health* Examines and quantifies dynamic processes in plant and fungal systems such as cell trafficking,growth of pollen tubes in model systems such as Neurospora Crassa and Arabidopsis* Discusses the dynamics of intracellular molecules for DNA repair and the regulation of cofilintransport using video analysis* Discusses software, system and database aspects of video bioinformatics by providing examples of5D cell tracking by FARSIGHT open source toolkit, a survey on available databases and software,biological processes for non-verbal communications and identification and retrieval of moth imagesThis unique text will be of great interest to researchers and graduate students of Electrical Engineering,Computer Science, Bioengineering, Cell Biology, Toxicology, Genetics, Genomics, Bioinformatics, ComputerVision and Pattern Recognition, Medical Image Analysis, and Cell Molecular and Developmental Biology.The large number of example applications will also appeal to application scientists and engineers.Dr. Bir Bhanu is Distinguished Professor of Electrical & C omputer Engineering, Interim Chair of theDepartment of Bioengineering, Cooperative Professor of Computer Science & Engineering, and MechanicalEngineering and the Director of the Center for Research in Intelligent Systems, at the University of California,Riverside, California, USA.Dr. Prue Talbot is Professor of Cell Biology & Neuroscience and Director of the Stem Cell Center and Core atthe University of California Riverside, California, USA. |
![]() ![]() You may like...
Fundamentals of Spatial Information…
Robert Laurini, Derek Thompson
Hardcover
R1,539
Discovery Miles 15 390
Big Data and Smart Service Systems
Xiwei Liu, Rangachari Anand, …
Hardcover
Artificial Intelligence and IoT - Smart…
Kalaiselvi Geetha Manoharan, Jawaharlal Arun Nehru, …
Hardcover
R5,605
Discovery Miles 56 050
Recent Advances in Technology Acceptance…
Mostafa Al-Emran, Khaled Shaalan
Hardcover
R5,645
Discovery Miles 56 450
Blockchain Technology and Computational…
Shahnawaz Khan, Mohammad Haider, …
Hardcover
R7,211
Discovery Miles 72 110
Computer Science Protecting Human…
Aleksander Byrski, Tadeusz Czachorski, …
Hardcover
R2,614
Discovery Miles 26 140
|