![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Applications of computing > Databases > Data mining
Graph data is powerful, thanks to its ability to model arbitrary relationship between objects and is encountered in a range of real-world applications in fields such as bioinformatics, traffic network, scientific collaboration, world wide web and social networks. Graph data mining is used to discover useful information and knowledge from graph data. The complications of nodes, links and the semi-structure form present challenges in terms of the computation tasks, e.g., node classification, link prediction, and graph classification. In this context, various advanced techniques, including graph embedding and graph neural networks, have recently been proposed to improve the performance of graph data mining. This book provides a state-of-the-art review of graph data mining methods. It addresses a current hot topic - the security of graph data mining - and proposes a series of detection methods to identify adversarial samples in graph data. In addition, it introduces readers to graph augmentation and subgraph networks to further enhance the models, i.e., improve their accuracy and robustness. Lastly, the book describes the applications of these advanced techniques in various scenarios, such as traffic networks, social and technical networks, and blockchains.
Recommender systems, software programs that learn from human behavior and make predictions of what products we are expected to appreciate and purchase, have become an integral part of our everyday life. They proliferate across electronic commerce around the globe and exist for virtually all sorts of consumable goods, such as books, movies, music, or clothes. At the same time, a new evolution on the Web has started to take shape, commonly known as the "Web 2.0" or the "Social Web" Consumer-generated media has become rife, social networks have emerged and are pulling significant shares of Web traffic. In line with these developments, novel information and knowledge artifacts have become readily available on the Web, created by the collective effort of millions of people. This textbook presents approaches to exploit the new Social Web fountain of knowledge, zeroing in first and foremost on two of those information artifacts, namely classification taxonomies and trust networks. These two are used to improve the performance of product-focused recommender systems: While classification taxonomies are appropriate means to fight the sparsity problem prevalent in many productive recommender systems, interpersonal trust ties - when used as proxies for interest similarity - are able to mitigate the recommenders' scalability problem.
Continuous improvements in data analysis and cloud computing have allowed more opportunities to develop systems with user-focused designs. This not only leads to higher success in day-to-day usage, but it increases the overall probability of technology adoption. Advancing Cloud Database Systems and Capacity Planning with Dynamic Applications is a key resource on the latest innovations in cloud database systems and their impact on the daily lives of people in modern society. Highlighting multidisciplinary studies on information storage and retrieval, big data architectures, and artificial intelligence, this publication is an ideal reference source for academicians, researchers, scientists, advanced level students, technology developers and IT officials.
Data clustering is a highly interdisciplinary field, the goal of which is to divide a set of objects into homogeneous groups such that objects in the same group are similar and objects in different groups are quite distinct. Thousands of theoretical papers and a number of books on data clustering have been published over the past 50 years. However, few books exist to teach people how to implement data clustering algorithms. This book was written for anyone who wants to implement or improve their data clustering algorithms. Using object-oriented design and programming techniques, Data Clustering in C++ exploits the commonalities of all data clustering algorithms to create a flexible set of reusable classes that simplifies the implementation of any data clustering algorithm. Readers can follow the development of the base data clustering classes and several popular data clustering algorithms. Additional topics such as data pre-processing, data visualization, cluster visualization, and cluster interpretation are briefly covered. This book is divided into three parts-- * Data Clustering and C++ Preliminaries: A review of basic concepts of data clustering, the unified modeling language, object-oriented programming in C++, and design patterns * A C++ Data Clustering Framework: The development of data clustering base classes * Data Clustering Algorithms: The implementation of several popular data clustering algorithms A key to learning a clustering algorithm is to implement and experiment the clustering algorithm. Complete listings of classes, examples, unit test cases, and GNU configuration files are included in the appendices of this book as well as in the CD-ROM of the book. The only requirements to compile the code are a modern C++ compiler and the Boost C++ libraries.
As urban congestion continues to be an ever increasing problem, routing in these settings has become an important area of operations research. This monograph provides cutting-edge research, utilizing the recent advances in technology, to quantify the value of dynamic, time-dependent information for advanced vehicle routing in city logistics. The methodology of traffic data collection is enhanced by GPS based data collection, resulting in a comprehensive number of travel time records. Data Mining is also applied to derive dynamic information models as required by time-dependent optimization. Finally, well-known approaches of vehicle routing are adapted in order to handle dynamic information models. This book interweaves the usually distinct areas of traffic data collection, information retrieval and time-dependent optimization by an integrated methodological approach, which refers to synergies of Data Mining and Operations Research techniques by example of city logistics applications. These procedures will help improve the reliability of logistics services in congested urban areas.
Perceiving complex multidimensional problems has proven to be a difficult task for people to overcome. However, introducing composite indicators into such problems allows the opportunity to reduce the problem's complexity. Emerging Trends in the Development and Application of Composite Indicators is an authoritative reference source for the latest scholarly research on the benefits and challenges presented by building composite indicators, and how these techniques promote optimized critical thinking. Highlighting various indicator types and quantitative methods, this book is ideally designed for developers, researchers, public officials, and upper-level students.
Handbook of Educational Data Mining (EDM) provides a thorough overview of the current state of knowledge in this area. The first part of the book includes nine surveys and tutorials on the principal data mining techniques that have been applied in education. The second part presents a set of 25 case studies that give a rich overview of the problems that EDM has addressed. Researchers at the Forefront of the Field Discuss Essential Topics and the Latest Advances With contributions by well-known researchers from a variety of fields, the book reflects the multidisciplinary nature of the EDM community. It brings the educational and data mining communities together, helping education experts understand what types of questions EDM can address and helping data miners understand what types of questions are important to educational design and educational decision making. Encouraging readers to integrate EDM into their research and practice, this timely handbook offers a broad, accessible treatment of essential EDM techniques and applications. It provides an excellent first step for newcomers to the EDM community and for active researchers to keep abreast of recent developments in the field.
Spatial trajectories have been bringing the unprecedented wealth to a variety of research communities. A spatial trajectory records the paths of a variety of moving objects, such as people who log their travel routes with GPS trajectories. The field of moving objects related research has become extremely active within the last few years, especially with all major database and data mining conferences and journals. "Computing with Spatial Trajectories" introduces the algorithms, technologies, and systems used to process, manage and understand existing spatial trajectories for different applications. This book also presents an overview on both fundamentals and the state-of-the-art research inspired by spatial trajectory data, as well as a special focus on trajectory pattern mining, spatio-temporal data mining and location-based social networks. Each chapter provides readers with a tutorial-style introduction to one important aspect of location trajectory computing, case studies and many valuable references to other relevant research work. "Computing with Spatial Trajectories" is designed as a reference or secondary text book for advanced-level students and researchers mainly focused on computer science and geography. Professionals working on spatial trajectory computing will also find this book very useful.
Data Mining is the science and technology of exploring large and complex bodies of data in order to discover useful patterns. It is extremely important because it enables modeling and knowledge extraction from abundant data availability. This book introduces soft computing methods extending the envelope of problems that data mining can solve efficiently. It presents practical soft-computing approaches in data mining and includes various real-world case studies with detailed results.
This edited volume presents the best chapters presented during the international conference on computer and applications ICCA'17 which was held in Dubai, United Arab Emirates in September 2017. Selected chapters present new advances in digital information, communications and multimedia. Authors from different countries show and discuss their findings, propose new approaches, compare them with the existing ones and include recommendations. They address all applications of computing including (but not limited to) connected health, information security, assistive technology, edutainment and serious games, education, grid computing, transportation, social computing, natural language processing, knowledge extraction and reasoning, Arabic apps, image and pattern processing, virtual reality, cloud computing, haptics, information security, robotics, networks algorithms, web engineering, big data analytics, ontology, constraints satisfaction, cryptography and steganography, Fuzzy logic, soft computing, neural networks, artificial intelligence, biometry and bio-informatics, embedded systems, computer graphics, algorithms and optimization, Internet of things and smart cities. The book can be used by researchers and practitioners to discover the recent trends in computer applications. It opens a new horizon for research discovery works locally and internationally.
The book reports on the latest advances and challenges of soft computing. Itgathers original scientific contributions written by top scientists in the fieldand covering theories, methods and applications in a number of research areas related to soft-computing, such as decision-making, probabilistic reasoning, image processing, control, neural networks and data analysis."
This book offers a timely report on key theories and applications of soft-computing. Written in honour of Professor Gaspar Mayor on his 70th birthday, it primarily focuses on areas related to his research, including fuzzy binary operators, aggregation functions, multi-distances, and fuzzy consensus/decision models. It also discusses a number of interesting applications such as the implementation of fuzzy mathematical morphology based on Mayor-Torrens t-norms. Importantly, the different chapters, authored by leading experts, present novel results and offer new perspectives on different aspects of Mayor's research. The book also includes an overview of evolutionary fuzzy systems, a topic that is not one of Mayor's main areas of interest, and a final chapter written by the Spanish pioneer in fuzzy logic, Professor E. Trillas. Computer and decision scientists, knowledge engineers and mathematicians alike will find here an authoritative overview of key soft-computing concepts and techniques.
Data mining is the process of extracting hidden patterns from data, and it's commonly used in business, bioinformatics, counter-terrorism, and, increasingly, in professional sports. First popularized in Michael Lewis' best-selling Moneyball: The Art of Winning An Unfair Game, it is has become an intrinsic part of all professional sports the world over, from baseball to cricket to soccer. While an industry has developed based on statistical analysis services for any given sport, or even for betting behavior analysis on these sports, no research-level book has considered the subject in any detail until now. Sports Data Mining brings together in one place the state of the art as it concerns an international array of sports: baseball, football, basketball, soccer, greyhound racing are all covered, and the authors (including Hsinchun Chen, one of the most esteemed and well-known experts in data mining in the world) present the latest research, developments, software available, and applications for each sport. They even examine the hidden patterns in gaming and wagering, along with the most common systems for wager analysis.
The importance of having ef cient and effective methods for data mining and kn- ledge discovery (DM&KD), to which the present book is devoted, grows every day and numerous such methods have been developed in recent decades. There exists a great variety of different settings for the main problem studied by data mining and knowledge discovery, and it seems that a very popular one is formulated in terms of binary attributes. In this setting, states of nature of the application area under consideration are described by Boolean vectors de ned on some attributes. That is, by data points de ned in the Boolean space of the attributes. It is postulated that there exists a partition of this space into two classes, which should be inferred as patterns on the attributes when only several data points are known, the so-called positive and negative training examples. The main problem in DM&KD is de ned as nding rules for recognizing (cl- sifying) new data points of unknown class, i. e. , deciding which of them are positive and which are negative. In other words, to infer the binary value of one more attribute, called the goal or class attribute. To solve this problem, some methods have been suggested which construct a Boolean function separating the two given sets of positive and negative training data points.
Recently, the pressure for fast processing and efficient storage of large data with complexrelations increased beyond the capability of traditional databases. Typical examples include iPhone applications, computer aided design - both electrical and mechanical, biochemistry applications, and incremental compilers. Serialization, which is sometimes used in such situations is notoriously tedious and error prone. In this book, Jiri Soukup and Petr Macha ek show in detail how to write programs which store their internal data automatically and transparently to disk. Together with special data structure libraries which treat relations among objects as first-class entities, and with a UML class-diagram generator, the core application code is much simplified. The benchmark chapter shows a typical example where persistent data is faster by the order of magnitude than with a traditional database, in both traversing and accessing the data. The authors explore and exploit advanced features of object-oriented languages in a depth hardly seen in print before. Yet, you as a reader need only a basic knowledge of C++, Java, C#, or Objective C. These languages are quite similar with respect to persistency, and the authors explain their differences where necessary. The book targets professional programmers working on any industry applications, it teaches you how to design your own persistent data or how to use the existing packages efficiently. Researchers in areas like language design, compiler construction, performance evaluation, and no-SQL applications will find a wealth of novel ideas and valuable implementation tips. Under http: //www.codefarms.com/bk, you will find a blog and other information, including a downloadable zip file with the sources of all the listings that are longer than just a few lines - ready to compile and run."
This book introduces readers to the field of conformance checking as a whole and outlines the fundamental relation between modelled and recorded behaviour. Conformance checking interrelates the modelled and recorded behaviour of a given process and provides techniques and methods for comparing and analysing observed instances of a process in the presence of a model, independent of the model's origin. Its goal is to provide an overview of the essential techniques and methods in this field at an intuitive level, together with precise formalisations of its underlying principles. The book is divided into three parts, that are meant to cover different perspectives of the field of conformance checking. Part I presents a comprehensive yet accessible overview of the essential concepts used to interrelate modelled and recorded behaviour. It also serves as a reference for assessing how conformance checking efforts could be applied in specific domains. Next, Part II provides readers with detailed insights into algorithms for conformance checking, including the most commonly used formal notions and their instantiation for specific analysis questions. Lastly, Part III highlights applications that help to make sense of conformance checking results, thereby providing a necessary next step to increase the value of a given process model. They help to interpret the outcomes of conformance checking and incorporate them by means of enhancement and repair techniques. Providing the core building blocks of conformance checking and describing its main applications, this book mainly addresses students specializing in business process management, researchers entering process mining and conformance checking for the first time, and advanced professionals whose work involves process evaluation, modelling and optimization.
Web usage mining is defined as the application of data mining technologies to online usage patterns as a way to better understand and serve the needs of web-based applications. Because the internet has become a central component in information sharing and commerce, having the ability to analyze user behavior on the web has become a critical component to a variety of industries. Web Usage Mining Techniques and Applications Across Industries addresses the systems and methodologies that enable organizations to predict web user behavior as a way to support website design and personalization of web-based services and commerce. Featuring perspectives from a variety of sectors, this publication is designed for use by IT specialists, business professionals, researchers, and graduate-level students interested in learning more about the latest concepts related to web-based information retrieval and mining.
This volume is aiming at a wide range of readers and researchers in the area of Big Data by presenting the recent advances in the fields of Big Data Analysis, as well as the techniques and tools used to analyze it. The book includes 10 distinct chapters providing a concise introduction to Big Data Analysis and recent Techniques and Environments for Big Data Analysis. It gives insight into how the expensive fitness evaluation of evolutionary learning can play a vital role in big data analysis by adopting Parallel, Grid, and Cloud computing environments.
This book provides a technical approach to a Business Resilience System with its Risk Atom and Processing Data Point based on fuzzy logic and cloud computation in real time. Its purpose and objectives define a clear set of expectations for Organizations and Enterprises so their network system and supply chain are totally resilient and protected against cyber-attacks, manmade threats, and natural disasters. These enterprises include financial, organizational, homeland security, and supply chain operations with multi-point manufacturing across the world. Market shares and marketing advantages are expected to result from the implementation of the system. The collected information and defined objectives form the basis to monitor and analyze the data through cloud computation, and will guarantee the success of their survivability's against any unexpected threats. This book will be useful for advanced undergraduate and graduate students in the field of computer engineering, engineers that work for manufacturing companies, business analysts in retail and e-Commerce, and those working in the defense industry, Information Security, and Information Technology.
Tools and methods from complex systems science can have a considerable impact on the way in which the quantitative assessment of economic and financial issues is approached, as discussed in this thesis. First it is shown that the self-organization of financial markets is a crucial factor in the understanding of their dynamics. In fact, using an agent-based approach, it is argued that financial markets' stylized facts appear only in the self-organized state. Secondly, the thesis points out the potential of so-called big data science for financial market modeling, investigating how web-driven data can yield a picture of market activities: it has been found that web query volumes anticipate trade volumes. As a third achievement, the metrics developed here for country competitiveness and product complexity is groundbreaking in comparison to mainstream theories of economic growth and technological development. A key element in assessing the intangible variables determining the success of countries in the present globalized economy is represented by the diversification of the productive basket of countries. The comparison between the level of complexity of a country's productive system and economic indicators such as the GDP per capita discloses its hidden growth potential.
Solving nonsmooth optimization (NSO) problems is critical in many practical applications and real-world modeling systems. The aim of this book is to survey various numerical methods for solving NSO problems and to provide an overview of the latest developments in the field. Experts from around the world share their perspectives on specific aspects of numerical NSO. The book is divided into four parts, the first of which considers general methods including subgradient, bundle and gradient sampling methods. In turn, the second focuses on methods that exploit the problem's special structure, e.g. algorithms for nonsmooth DC programming, VU decomposition techniques, and algorithms for minimax and piecewise differentiable problems. The third part considers methods for special problems like multiobjective and mixed integer NSO, and problems involving inexact data, while the last part highlights the latest advancements in derivative-free NSO. Given its scope, the book is ideal for students attending courses on numerical nonsmooth optimization, for lecturers who teach optimization courses, and for practitioners who apply nonsmooth optimization methods in engineering, artificial intelligence, machine learning, and business. Furthermore, it can serve as a reference text for experts dealing with nonsmooth optimization.
This book presents innovative work in Climate Informatics, a new field that reflects the application of data mining methods to climate science, and shows where this new and fast growing field is headed. Given its interdisciplinary nature, Climate Informatics offers insights, tools and methods that are increasingly needed in order to understand the climate system, an aspect which in turn has become crucial because of the threat of climate change. There has been a veritable explosion in the amount of data produced by satellites, environmental sensors and climate models that monitor, measure and forecast the earth system. In order to meaningfully pursue knowledge discovery on the basis of such voluminous and diverse datasets, it is necessary to apply machine learning methods, and Climate Informatics lies at the intersection of machine learning and climate science. This book grew out of the fourth workshop on Climate Informatics held in Boulder, Colorado in Sep. 2014. |
![]() ![]() You may like...
Thermal Control Thin Films - Spacecraft…
Jia-Wen Qiu, Yu-Dong Feng, …
Hardcover
R3,437
Discovery Miles 34 370
Information Support and Aircraft Flight…
Oleg Ivanovich Zavalishin, Anatoly Nikolaevich Korotonoshko, …
Hardcover
R4,316
Discovery Miles 43 160
Signal Processing of Airborne Radar…
Vereshchagin A.V., Zatuchny D.A., …
Hardcover
R3,809
Discovery Miles 38 090
Cases on Modern Computer Systems in…
Tetiana Shmelova, Yuliya Sikirda, …
Hardcover
R6,357
Discovery Miles 63 570
China Satellite Navigation Conference…
Jiadong Sun, Jingnan Liu, …
Hardcover
R6,013
Discovery Miles 60 130
Electro-Mechanical Actuators for the…
Mirko Mazzoleni, Gianpietro Di Rito, …
Hardcover
R5,025
Discovery Miles 50 250
|