![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Applications of computing > Databases > Data mining
NVIDIA's Full-Color Guide to Deep Learning: All You Need to Get Started and Get Results "To enable everyone to be part of this historic revolution requires the democratization of AI knowledge and resources. This book is timely and relevant towards accomplishing these lofty goals." -- From the foreword by Dr. Anima Anandkumar, Bren Professor, Caltech, and Director of ML Research, NVIDIA "Ekman uses a learning technique that in our experience has proven pivotal to success-asking the reader to think about using DL techniques in practice. His straightforward approach is refreshing, and he permits the reader to dream, just a bit, about where DL may yet take us." -- From the foreword by Dr. Craig Clawson, Director, NVIDIA Deep Learning Institute Deep learning (DL) is a key component of today's exciting advances in machine learning and artificial intelligence. Learning Deep Learning is a complete guide to DL. Illuminating both the core concepts and the hands-on programming techniques needed to succeed, this book is ideal for developers, data scientists, analysts, and others--including those with no prior machine learning or statistics experience. After introducing the essential building blocks of deep neural networks, such as artificial neurons and fully connected, convolutional, and recurrent layers, Magnus Ekman shows how to use them to build advanced architectures, including the Transformer. He describes how these concepts are used to build modern networks for computer vision and natural language processing (NLP), including Mask R-CNN, GPT, and BERT. And he explains how a natural language translator and a system generating natural language descriptions of images. Throughout, Ekman provides concise, well-annotated code examples using TensorFlow with Keras. Corresponding PyTorch examples are provided online, and the book thereby covers the two dominating Python libraries for DL used in industry and academia. He concludes with an introduction to neural architecture search (NAS), exploring important ethical issues and providing resources for further learning. Explore and master core concepts: perceptrons, gradient-based learning, sigmoid neurons, and back propagation See how DL frameworks make it easier to develop more complicated and useful neural networks Discover how convolutional neural networks (CNNs) revolutionize image classification and analysis Apply recurrent neural networks (RNNs) and long short-term memory (LSTM) to text and other variable-length sequences Master NLP with sequence-to-sequence networks and the Transformer architecture Build applications for natural language translation and image captioning NVIDIA's invention of the GPU sparked the PC gaming market. The company's pioneering work in accelerated computing--a supercharged form of computing at the intersection of computer graphics, high-performance computing, and AI--is reshaping trillion-dollar industries, such as transportation, healthcare, and manufacturing, and fueling the growth of many others. Register your book for convenient access to downloads, updates, and/or corrections as they become available. See inside book for details.
This edited book covers recent advances of techniques, methods and tools treating the problem of learning from data streams generated by evolving non-stationary processes. The goal is to discuss and overview the advanced techniques, methods and tools that are dedicated to manage, exploit and interpret data streams in non-stationary environments. The book includes the required notions, definitions, and background to understand the problem of learning from data streams in non-stationary environments and synthesizes the state-of-the-art in the domain, discussing advanced aspects and concepts and presenting open problems and future challenges in this field. Provides multiple examples to facilitate the understanding data streams in non-stationary environments; Presents several application cases to show how the methods solve different real world problems; Discusses the links between methods to help stimulate new research and application directions.
This open access book describes the results of natural language processing and machine learning methods applied to clinical text from electronic patient records. It is divided into twelve chapters. Chapters 1-4 discuss the history and background of the original paper-based patient records, their purpose, and how they are written and structured. These initial chapters do not require any technical or medical background knowledge. The remaining eight chapters are more technical in nature and describe various medical classifications and terminologies such as ICD diagnosis codes, SNOMED CT, MeSH, UMLS, and ATC. Chapters 5-10 cover basic tools for natural language processing and information retrieval, and how to apply them to clinical text. The difference between rule-based and machine learning-based methods, as well as between supervised and unsupervised machine learning methods, are also explained. Next, ethical concerns regarding the use of sensitive patient records for research purposes are discussed, including methods for de-identifying electronic patient records and safely storing patient records. The book's closing chapters present a number of applications in clinical text mining and summarise the lessons learned from the previous chapters. The book provides a comprehensive overview of technical issues arising in clinical text mining, and offers a valuable guide for advanced students in health informatics, computational linguistics, and information retrieval, and for researchers entering these fields.
The post-genomic revolution is witnessing the generation of petabytes of data annually, with deep implications ranging across evolutionary theory, developmental biology, agriculture, and disease processes. "Data Mining for Systems Biology: Methods and Protocols," surveys and demonstrates the science and technology of converting an unprecedented data deluge to new knowledge and biological insight. The volume is organized around two overlapping themes, network inference and functional inference. Written in the highly successful "Methods in Molecular Biology " series format, chapters include introductions to their respective topics, lists of the necessary materials and reagents, step-by-step, readily reproducible protocols, and key tips on troubleshooting and avoiding known pitfalls. Authoritative and practical, "Data Mining for Systems Biology: Methods and Protocols" also seeks to aid researchers in the further development of databases, mining and visualization systems that are central to the paradigm altering discoveries being made with increasing frequency."
This book addresses different methods and techniques of integration for enhancing the overall goal of data mining. The book is a collection of high-quality peer-reviewed research papers presented in the Sixth International Conference on Computational Intelligence in Data Mining (ICCIDM 2021) held at Aditya Institute of Technology and Management, Tekkali, Andhra Pradesh, India, during December 11-12, 2021. The book addresses the difficulties and challenges for the seamless integration of two core disciplines of computer science, i.e., computational intelligence and data mining. The book helps to disseminate the knowledge about some innovative, active research directions in the field of data mining, machine and computational intelligence, along with some current issues and applications of related topics.
The latest inventions in internet technology influence most of business and daily activities. Internet security, internet data management, web search, data grids, cloud computing, and web-based applications play vital roles, especially in business and industry, as more transactions go online and mobile. Issues related to ubiquitous computing are becoming critical. Internet technology and data engineering should reinforce efficiency and effectiveness of business processes. These technologies should help people make better and more accurate decisions by presenting necessary information and possible consequences for the decisions. Intelligent information systems should help us better understand and manage information with ubiquitous data repository and cloud computing. This book is a compilation of some recent research findings in Internet Technology and Data Engineering. This book provides state-of-the-art accounts in computational algorithms/tools, database management and database technologies, intelligent information systems, data engineering applications, internet security, internet data management, web search, data grids, cloud computing, web-based application, and other related topics.
Edge computing is quickly becoming an important technology throughout a number of fields as businesses and industries alike embrace the benefits it can have in their companies. The streamlining of data is crucial for the development and evolution of businesses in order to keep up with competition and improve functions overall. In order to appropriately utilize edge computing to its full potential, further study is required to examine the potential pitfalls and opportunities of this innovative technology. The Research Anthology on Edge Computing Protocols, Applications, and Integration establishes critical research on the current uses, innovations, and challenges of edge computing across disciplines. The text highlights the history of edge computing and how it has been adapted over time to improve industries. Covering a range of topics such as bandwidth, data centers, and security, this major reference work is ideal for industry professionals, computer scientists, engineers, practitioners, researchers, academicians, scholars, instructors, and students.
Formal specifications are an important tool for the construction, verification and analysis of systems, since without it is hardly possible to explain whether a system worked correctly or showed an expected behavior. This book proposes the use of representation theorems as a means to develop an understanding of all models of a specification in order to exclude possible unintended models, demonstrating the general methodology with representation theorems for applications in qualitative spatial reasoning, data stream processing, and belief revision. For qualitative spatial reasoning, it develops a model of spatial relatedness that captures the scaling context with hierarchical partitions of a spatial domain, and axiomatically characterizes the resulting relations. It also shows that various important properties of stream processing, such as prefix-determinedness or various factorization properties can be axiomatized, and that the axioms are fulfilled by natural classes of stream functions. The third example is belief revision, which is concerned with the revision of knowledge bases under new, potentially incompatible information. In this context, the book considers a subclass of revision operators, namely the class of reinterpretation operators, and characterizes them axiomatically. A characteristic property of reinterpretation operators is that of dissolving potential inconsistencies by reinterpreting symbols of the knowledge base. Intended for researchers in theoretical computer science or one of the above application domains, the book presents results that demonstrate the use of representation theorems for the design and evaluation of formal specifications, and provide the basis for future application-development kits that support application designers with automatically built representations.
Making use of data is not anymore a niche project but central to almost every project. With access to massive compute resources and vast amounts of data, it seems at least in principle possible to solve any problem. However, successful data science projects result from the intelligent application of: human intuition in combination with computational power; sound background knowledge with computer-aided modelling; and critical reflection of the obtained insights and results. Substantially updating the previous edition, then entitled Guide to Intelligent Data Analysis, this core textbook continues to provide a hands-on instructional approach to many data science techniques, and explains how these are used to solve real world problems. The work balances the practical aspects of applying and using data science techniques with the theoretical and algorithmic underpinnings from mathematics and statistics. Major updates on techniques and subject coverage (including deep learning) are included. Topics and features: guides the reader through the process of data science, following the interdependent steps of project understanding, data understanding, data blending and transformation, modeling, as well as deployment and monitoring; includes numerous examples using the open source KNIME Analytics Platform, together with an introductory appendix; provides a review of the basics of classical statistics that support and justify many data analysis methods, and a glossary of statistical terms; integrates illustrations and case-study-style examples to support pedagogical exposition; supplies further tools and information at an associated website. This practical and systematic textbook/reference is a "need-to-have" tool for graduate and advanced undergraduate students and essential reading for all professionals who face data science problems. Moreover, it is a "need to use, need to keep" resource following one's exploration of the subject.
Biomarker discovery is an important area of biomedical research that may lead to significant breakthroughs in disease analysis and targeted therapy. Biomarkers are biological entities whose alterations are measurable and are characteristic of a particular biological condition. Discovering, managing, and interpreting knowledge of new biomarkers are challenging and attractive problems in the emerging field of biomedical informatics. This volume is a collection of state-of-the-art research into the application of data mining to the discovery and analysis of new biomarkers. Presenting new results, models and algorithms, the included contributions focus on biomarker data integration, information retrieval methods, and statistical machine learning techniques. This volume is intended for students, and researchers in bioinformatics, proteomics, and genomics, as well engineers and applied scientists interested in the interdisciplinary application of data mining techniques.
Nearly everyone knows K-means algorithm in the fields of data mining and business intelligence. But the ever-emerging data with extremely complicated characteristics bring new challenges to this "old" algorithm. This book addresses these challenges and makes novel contributions in establishing theoretical frameworks for K-means distances and K-means based consensus clustering, identifying the "dangerous" uniform effect and zero-value dilemma of K-means, adapting right measures for cluster validity, and integrating K-means with SVMs for rare class analysis. This book not only enriches the clustering and optimization theories, but also provides good guidance for the practical use of K-means, especially for important tasks such as network intrusion detection and credit fraud prediction. The thesis on which this book is based has won the "2010 National Excellent Doctoral Dissertation Award", the highest honor for not more than 100 PhD theses per year in China.
Big data is a well-trafficked subject in recent IT discourse and does not lack for current research. In fact, there is such a surfeit of material related to big data-and so much of it of questionably reliability, thanks to the high-gloss efforts of savvy tech-marketing gurus-that it can, at times, be difficult for a serious academician to navigate. The Handbook of Research on Trends and Future Directions in Big Data and Web Intelligence cuts through the haze of glitz and pomp surrounding big data and offers a simple, straightforward reference-source of practical academic utility. Covering such topics as cloud computing, parallel computing, natural language processing, and personalized medicine, this volume presents an overview of current research, insight into recent advances, and gaps in the literature indicative of opportunities for future inquiry and is targeted toward a broad, interdisciplinary audience of students, academics, researchers, and professionals in fields of IT, networking, and data-analytics.
Social media data contains our communication and online sharing, mirroring our daily life. This book looks at how we can use and what we can discover from such big data: Basic knowledge (data & challenges) on social media analytics Clustering as a fundamental technique for unsupervised knowledge discovery and data mining A class of neural inspired algorithms, based on adaptive resonance theory (ART), tackling challenges in big social media data clustering Step-by-step practices of developing unsupervised machine learning algorithms for real-world applications in social media domain Adaptive Resonance Theory in Social Media Data Clustering stands on the fundamental breakthrough in cognitive and neural theory, i.e. adaptive resonance theory, which simulates how a brain processes information to perform memory, learning, recognition, and prediction. It presents initiatives on the mathematical demonstration of ART's learning mechanisms in clustering, and illustrates how to extend the base ART model to handle the complexity and characteristics of social media data and perform associative analytical tasks. Both cutting-edge research and real-world practices on machine learning and social media analytics are included in the book and if you wish to learn the answers to the following questions, this book is for you: How to process big streams of multimedia data? How to analyze social networks with heterogeneous data? How to understand a user's interests by learning from online posts and behaviors? How to create a personalized search engine by automatically indexing and searching multimodal information resources? .
Analyzing Social Media Networks with NodeXL: Insights from a Connected World, Second Edition, provides readers with a thorough, practical and updated guide to NodeXL, the open-source social network analysis (SNA) plug-in for use with Excel. The book analyzes social media, provides a NodeXL tutorial, and presents network analysis case studies, all of which are revised to reflect the latest developments. Sections cover history and concepts, mapping and modeling, the detailed operation of NodeXL, and case studies, including e-mail, Twitter, Facebook, Flickr and YouTube. In addition, there are descriptions of each system and types of analysis for identifying people, documents, groups and events. This book is perfect for use as a course text in social network analysis or as a guide for practicing NodeXL users.
This work provides an innovative look at the use of open data for extracting information to detect and prevent crime, and also explores the link between terrorism and organized crime. In counter-terrorism and other forms of crime prevention, foresight about potential threats is vitally important and this information is increasingly available via electronic data sources such as social media communications. However, the amount and quality of these sources is varied, and researchers and law enforcement need guidance about when and how to extract useful information from them. The emergence of these crime threats, such as communication between organized crime networks and radicalization towards terrorism, is driven by a combination of political, economic, social, technological, legal and environmental factors. The contributions to this volume represent a major step by researchers to systematically collect, filter, interpret, and use the information available. For the purposes of this book, the only data sources used are publicly available sources which can be accessed legally and ethically. This work will be of interest to researchers in criminology and criminal justice, particularly in police science, organized crime, counter-terrorism and crime science. It will also be of interest to those in related fields such as applications of computer science and data mining, public policy, and business intelligence.
A pathbreaking history of art that uses digital research and economic tools to reveal enduring inequities in the formation of the art historical canon Painting by Numbers presents a groundbreaking blend of art historical and social scientific methods to chart, for the first time, the sheer scale of nineteenth-century artistic production. With new quantitative evidence for more than five hundred thousand works of art, Diana Seave Greenwald provides fresh insights into the nineteenth century, and the extent to which art historians have focused on a limited-and potentially biased-sample of artwork from that time. She addresses long-standing questions about the effects of industrialization, gender, and empire on the art world, and she models more expansive approaches for studying art history in the age of the digital humanities. Examining art in France, the United States, and the United Kingdom, Greenwald features datasets created from indices and exhibition catalogs that-to date-have been used primarily as finding aids. From this body of information, she reveals the importance of access to the countryside for painters showing images of nature at the Paris Salon, the ways in which time-consuming domestic responsibilities pushed women artists in the United States to work in lower-prestige genres, and how images of empire were largely absent from the walls of London's Royal Academy at the height of British imperial power. Ultimately, Greenwald considers how many works may have been excluded from art historical inquiry and shows how data can help reintegrate them into the history of art, even after such pieces have disappeared or faded into obscurity. Upending traditional perspectives on the art historical canon, Painting by Numbers offers an innovative look at the nineteenth-century art world and its legacy.
This book provides an overview of data mining methods demonstrated by software. Knowledge management involves application of human knowledge (epistemology) with the technological advances of our current society (computer systems) and big data, both in terms of collecting data and in analyzing it. We see three types of analytic tools. Descriptive analytics focus on reports of what has happened. Predictive analytics extend statistical and/or artificial intelligence to provide forecasting capability. It also includes classification modeling. Diagnostic analytics can apply analysis to sensor input to direct control systems automatically. Prescriptive analytics applies quantitative models to optimize systems, or at least to identify improved systems. Data mining includes descriptive and predictive modeling. Operations research includes all three. This book focuses on descriptive analytics. The book seeks to provide simple explanations and demonstration of some descriptive tools. This second edition provides more examples of big data impact, updates the content on visualization, clarifies some points, and expands coverage of association rules and cluster analysis. Chapter 1 gives an overview in the context of knowledge management. Chapter 2 discusses some basic software support to data visualization. Chapter 3 covers fundamentals of market basket analysis, and Chapter 4 provides demonstration of RFM modeling, a basic marketing data mining tool. Chapter 5 demonstrates association rule mining. Chapter 6 is a more in-depth coverage of cluster analysis. Chapter 7 discusses link analysis. Models are demonstrated using business related data. The style of the book is intended to be descriptive, seeking to explain how methods work, with some citations, but without deep scholarly reference. The data sets and software are all selected for widespread availability and access by any reader with computer links.
The need for both organizations and government agencies to generate, collect, and utilize data in public and private sector activities is rapidly increasing, placing importance on the growth of data mining applications and tools. Data Mining in Public and Private Sectors: Organizational and Government Applications explores the manifestation of data mining and how it can be enhanced at various levels of management. This innovative publication provides relevant theoretical frameworks and the latest empirical research findings useful to governmental agencies, practicing managers, and academicians.
The three volume set provides a systematic overview of theories and technique on social network analysis.Volume 2 of the set mainly focuses on the formation and interaction of group behaviors. Users' behavior analysis, sentiment analysis, influence analysis and collective aggregation are discussed in detail as well. It is an essential reference for scientist and professionals in computer science.
This book presents the complex topic of using computational intelligence for pattern recognition in a straightforward and applicable way, using Matlab to illustrate topics and concepts. The author covers computational intelligence tools like particle swarm optimization, bacterial foraging, simulated annealing, genetic algorithm, and artificial neural networks. The Matlab based illustrations along with the code are given for every topic. Readers get a quick basic understanding of various pattern recognition techniques using only the required depth in math. The Matlab program and algorithm are given along with the running text, providing clarity and usefulness of the various techniques. Presents pattern recognition and the computational intelligence using Matlab; Includes mixtures of theory, math, and algorithms, letting readers understand the concepts quickly; Outlines an array of classifiers, various regression models, statistical tests and the techniques for pattern recognition using computational intelligence.
This book studies mathematical theories of machine learning. The first part of the book explores the optimality and adaptivity of choosing step sizes of gradient descent for escaping strict saddle points in non-convex optimization problems. In the second part, the authors propose algorithms to find local minima in nonconvex optimization and to obtain global minima in some degree from the Newton Second Law without friction. In the third part, the authors study the problem of subspace clustering with noisy and missing data, which is a problem well-motivated by practical applications data subject to stochastic Gaussian noise and/or incomplete data with uniformly missing entries. In the last part, the authors introduce an novel VAR model with Elastic-Net regularization and its equivalent Bayesian model allowing for both a stable sparsity and a group selection.
This book presents the proceedings of the Conference on Algorithms and Applications (ALAP 2018), which focuses on various areas of computing, like distributed systems and security, big data and analytics and very-large-scale integration (VLSI) design. The book provides solutions to a broad class of problems in diverse areas of algorithms in our daily lives in a world designed for, and increasingly controlled by algorithms. Written by eminent personalities from academia and industry, the papers included offer insights from a number of perspectives, providing an overview of the state of the art in the field. The book consists of invited talks by respected speakers, papers presented in technical sessions, and tutorials to offer ideas, results, work-in-progress and experiences of various algorithmic aspects of computational science and engineering.
Intelligent Data Mining and Fusion Systems in Agriculture presents methods of computational intelligence and data fusion that have applications in agriculture for the non-destructive testing of agricultural products and crop condition monitoring. Sections cover the combination of sensors with artificial intelligence architectures in precision agriculture, including algorithms, bio-inspired hierarchical neural maps, and novelty detection algorithms capable of detecting sudden changes in different conditions. This book offers advanced students and entry-level professionals in agricultural science and engineering, geography and geoinformation science an in-depth overview of the connection between decision-making in agricultural operations and the decision support features offered by advanced computational intelligence algorithms.
This textbook grew out of notes for the ECE143 Programming for Data Analysis class that the author has been teaching at University of California, San Diego, which is a requirement for both graduate and undergraduate degrees in Machine Learning and Data Science. This book is ideal for readers with some Python programming experience. The book covers key language concepts that must be understood to program effectively, especially for data analysis applications. Certain low-level language features are discussed in detail, especially Python memory management and data structures. Using Python effectively means taking advantage of its vast ecosystem. The book discusses Python package management and how to use third-party modules as well as how to structure your own Python modules. The section on object-oriented programming explains features of the language that facilitate common programming patterns. After developing the key Python language features, the book moves on to third-party modules that are foundational for effective data analysis, starting with Numpy. The book develops key Numpy concepts and discusses internal Numpy array data structures and memory usage. Then, the author moves onto Pandas and details its many features for data processing and alignment. Because strong visualizations are important for communicating data analysis, key modules such as Matplotlib are developed in detail, along with web-based options such as Bokeh, Holoviews, Altair, and Plotly. The text is sprinkled with many tricks-of-the-trade that help avoid common pitfalls. The author explains the internal logic embodied in the Python language so that readers can get into the Python mindset and make better design choices in their codes, which is especially helpful for newcomers to both Python and data analysis. To get the most out of this book, open a Python interpreter and type along with the many code samples.
The last decade has witnessed the rise of big data in game development as the increasing proliferation of Internet-enabled gaming devices has made it easier than ever before to collect large amounts of player-related data. At the same time, the emergence of new business models and the diversification of the player base have exposed a broader potential audience, which attaches great importance to being able to tailor game experiences to a wide range of preferences and skill levels. This, in turn, has led to a growing interest in data mining techniques, as they offer new opportunities for deriving actionable insights to inform game design, to ensure customer satisfaction, to maximize revenues, and to drive technical innovation. By now, data mining and analytics have become vital components of game development. The amount of work being done in this area nowadays makes this an ideal time to put together a book on this subject. Data Analytics Applications in Gaming and Entertainment seeks to provide a cross section of current data analytics applications in game production. It is intended as a companion for practitioners, academic researchers, and students seeking knowledge on the latest practices in game data mining. The chapters have been chosen in such a way as to cover a wide range of topics and to provide readers with a glimpse at the variety of applications of data mining in gaming. A total of 25 authors from industry and academia have contributed 12 chapters covering topics such as player profiling, approaches for analyzing player communities and their social structures, matchmaking, churn prediction and customer lifetime value estimation, communication of analytical results, and visual approaches to game analytics. This book's perspectives and concepts will spark heightened interest in game analytics and foment innovative ideas that will advance the exciting field of online gaming and entertainment. |
![]() ![]() You may like...
Information Technology in Disaster Risk…
Yuko Murayama, Dimiter Velev, …
Hardcover
R2,915
Discovery Miles 29 150
Literate Apprenticeships - The Emergence…
Kenneth Reeder, Jon Shapiro, …
Hardcover
R2,213
Discovery Miles 22 130
Dynamics of Coupled Map Lattices and of…
Jean-Rene Chazottes, Bastien Fernandez
Hardcover
R3,089
Discovery Miles 30 890
Beyond Principles and Parameters…
Kyle Johnson, I. G. Roberts
Hardcover
R3,000
Discovery Miles 30 000
|