![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Databases
The rapid advancement of semantic web technologies, along with the fact that they are at various levels of maturity, has left many practitioners confused about the current state of these technologies. Focusing on the most mature technologies, Applied Semantic Web Technologies integrates theory with case studies to illustrate the history, current state, and future direction of the semantic web. It maintains an emphasis on real-world applications and examines the technical and practical issues related to the use of semantic technologies in intelligent information management. The book starts with an introduction to the fundamentals-reviewing ontology basics, ontology languages, and research related to ontology alignment, mediation, and mapping. Next, it covers ontology engineering issues and presents a collaborative ontology engineering tool that is an extension of the Semantic MediaWiki. Unveiling a novel approach to data and knowledge engineering, the text: Introduces cutting-edge taxonomy-aware algorithms Examines semantics-based service composition in transport logistics Offers ontology alignment tools that use information visualization techniques Explains how to enrich the representation of entity semantics in an ontology Addresses challenges in tackling the content creation bottleneck Using case studies, the book provides authoritative insights and highlights valuable lessons learned by the authors-information systems veterans with decades of experience. They explain how to create social ontologies and present examples of the application of semantic technologies in building automation, logistics, ontology-driven business process intelligence, decision making, and energy efficiency in smart homes.
In 2013, the International Conference on Advance Information Systems Engineering (CAiSE) turns 25. Initially launched in 1989, for all these years the conference has provided a broad forum for researchers working in the area of Information Systems Engineering. To reflect on the work done so far and to examine prospects for future work, the CAiSE Steering Committee decided to present a selection of seminal papers published for the conference during these years and to ask their authors, all prominent researchers in the field, to comment on their work and how it has developed over the years. The scope of the papers selected covers a broad range of topics related to modeling and designing information systems, collecting and managing requirements, and with special attention to how information systems are engineered towards their final development and deployment as software components.With this approach, the book provides not only a historical analysis on how information systems engineering evolved over the years, but also a fascinating social network analysis of the research community. Additionally, many inspiring ideas for future research and new perspectives in this area are sparked by the intriguing comments of the renowned authors.
Updated new edition of Ralph Kimball's groundbreaking book on dimensional modeling for data warehousing and business intelligence The first edition of Ralph Kimball's "The Data Warehouse Toolkit" introduced the industry to dimensional modeling, and now his books are considered the most authoritative guides in this space. This new third edition is a complete library of updated dimensional modeling techniques, the most comprehensive collection ever. It covers new and enhanced star schema dimensional modeling patterns, adds two new chapters on ETL techniques, includes new and expanded business matrices for 12 case studies, and more.Authored by Ralph Kimball and Margy Ross, known worldwide as educators, consultants, and influential thought leaders in data warehousing and business intelligenceBegins with fundamental design recommendations and progresses through increasingly complex scenariosPresents unique modeling techniques for business applications such as inventory management, procurement, invoicing, accounting, customer relationship management, big data analytics, and moreDraws real-world case studies from a variety of industries, including retail sales, financial services, telecommunications, education, health care, insurance, e-commerce, and more Design dimensional databases that are easy to understand and provide fast query response with "The Data Warehouse Toolkit: The Definitive Guide to Dimensional Modeling, 3rd Edition."
Video on Demand Systems brings together in one place important contributions and up-to-date research results in this fast moving area. Video on Demand Systems serves as an excellent reference, providing insight into some of the most challenging research issues in the field.
This proceedings volume brings together the results of a corporate discussion on research, academic teaching and education in the field of business and economics in the context of globalization. The contributions examine leadership and sustainability, quality and governance and the internationalization of higher education. With a particular focus on business education and business schools, the book discusses the labor market and modernization as well as contemporary trends and challenges. By including both academic papers and contributions from industry, it forges research links between academia, business and industry.
Data mining involves the non-trivial extraction of implicit, previously unknown, and potentially useful information from databases. Genetic Programming (GP) and Inductive Logic Programming (ILP) are two of the approaches for data mining. This book first sets the necessary backgrounds for the reader, including an overview of data mining, evolutionary algorithms and inductive logic programming. It then describes a framework, called GGP (Generic Genetic Programming), that integrates GP and ILP based on a formalism of logic grammars. The formalism is powerful enough to represent context- sensitive information and domain-dependent knowledge. This knowledge can be used to accelerate the learning speed and/or improve the quality of the knowledge induced. A grammar-based genetic programming system called LOGENPRO (The LOGic grammar based GENetic PROgramming system) is detailed and tested on many problems in data mining. It is found that LOGENPRO outperforms some ILP systems. We have also illustrated how to apply LOGENPRO to emulate Automatically Defined Functions (ADFs) to discover problem representation primitives automatically. By employing various knowledge about the problem being solved, LOGENPRO can find a solution much faster than ADFs and the computation required by LOGENPRO is much smaller than that of ADFs. Moreover, LOGENPRO can emulate the effects of Strongly Type Genetic Programming and ADFs simultaneously and effortlessly. Data Mining Using Grammar Based Genetic Programming and Applications is appropriate for researchers, practitioners and clinicians interested in genetic programming, data mining, and the extraction of data from databases.
Method Engineering focuses on the design, construction and evaluation of methods, techniques and support tools for information systems development It addresses a number of important topics, including: method representation formalisms; meta-modelling; situational methods; contingency approaches; system development practices of method engineering; terminology and reference models; ontologies; usability and experience reports; and organisational support and impact.
"JDBC Metadata, MySQL, and Oracle Recipes" is the only book that focuses on metadata or annotation-based code recipes for JDBC API for use with Oracle and MySQL. It continues where the authors other book, "JDBC Recipes: A Problem-Solution Approach," leaves off. This edition is also a Java EE 5-compliant book, perfect for lightweight Java database development. And it provides cut-and-paste code templates that can be immediately customized and applied in each developer's application development.
Research Directions in Data and Applications Security describes original research results and innovative practical developments, all focused on maintaining security and privacy in database systems and applications that pervade cyberspace. The areas of coverage include: -Role-Based Access Control;
Data Mining is the science and technology of exploring large and complex bodies of data in order to discover useful patterns. It is extremely important because it enables modeling and knowledge extraction from abundant data availability. This book introduces soft computing methods extending the envelope of problems that data mining can solve efficiently. It presents practical soft-computing approaches in data mining and includes various real-world case studies with detailed results.
This book introduces advanced semantic web technologies, illustrating their utility and highlighting their implementation in biological, medical, and clinical scenarios. It covers topics ranging from database, ontology, and visualization to semantic web services and workflows. The volume also details the factors impacting on the establishment of the semantic web in life science and the legal challenges that will impact on its proliferation.
Hybrid Intelligent Systems for Information Retrieval covers three areas along with the introduction to Intelligent IR, i.e., Optimal Information Retrieval Using Evolutionary Approaches, Semantic Search for Web Information Retrieval, and Natural Language Processing for Information Retrieval. * Talks about the design, implementation, and performance issues of the hybrid intelligent information retrieval system in one book * Gives a clear insight into challenges and issues in designing a hybrid information retrieval system * Includes case studies on structured and unstructured data for hybrid intelligent information retrieval * Provides research directions for the design and development of intelligent search engines This book is aimed primarily at graduates and researchers in the information retrieval domain.
The book reports on the latest advances and challenges of soft computing. Itgathers original scientific contributions written by top scientists in the fieldand covering theories, methods and applications in a number of research areas related to soft-computing, such as decision-making, probabilistic reasoning, image processing, control, neural networks and data analysis."
Data clustering is a highly interdisciplinary field, the goal of which is to divide a set of objects into homogeneous groups such that objects in the same group are similar and objects in different groups are quite distinct. Thousands of theoretical papers and a number of books on data clustering have been published over the past 50 years. However, few books exist to teach people how to implement data clustering algorithms. This book was written for anyone who wants to implement or improve their data clustering algorithms. Using object-oriented design and programming techniques, Data Clustering in C++ exploits the commonalities of all data clustering algorithms to create a flexible set of reusable classes that simplifies the implementation of any data clustering algorithm. Readers can follow the development of the base data clustering classes and several popular data clustering algorithms. Additional topics such as data pre-processing, data visualization, cluster visualization, and cluster interpretation are briefly covered. This book is divided into three parts-- * Data Clustering and C++ Preliminaries: A review of basic concepts of data clustering, the unified modeling language, object-oriented programming in C++, and design patterns * A C++ Data Clustering Framework: The development of data clustering base classes * Data Clustering Algorithms: The implementation of several popular data clustering algorithms A key to learning a clustering algorithm is to implement and experiment the clustering algorithm. Complete listings of classes, examples, unit test cases, and GNU configuration files are included in the appendices of this book as well as in the CD-ROM of the book. The only requirements to compile the code are a modern C++ compiler and the Boost C++ libraries.
Mining Very Large Databases with Parallel Processing addresses the problem of large-scale data mining. It is an interdisciplinary text, describing advances in the integration of three computer science areas, namely intelligent' (machine learning-based) data mining techniques, relational databases and parallel processing. The basic idea is to use concepts and techniques of the latter two areas - particularly parallel processing - to speed up and scale up data mining algorithms. The book is divided into three parts. The first part presents a comprehensive review of intelligent data mining techniques such as rule induction, instance-based learning, neural networks and genetic algorithms. Likewise, the second part presents a comprehensive review of parallel processing and parallel databases. Each of these parts includes an overview of commercially-available, state-of-the-art tools. The third part deals with the application of parallel processing to data mining. The emphasis is on finding generic, cost-effective solutions for realistic data volumes. Two parallel computational environments are discussed, the first excluding the use of commercial-strength DBMS, and the second using parallel DBMS servers. It is assumed that the reader has a knowledge roughly equivalent to a first degree (BSc) in accurate sciences, so that (s)he is reasonably familiar with basic concepts of statistics and computer science. The primary audience for Mining Very Large Databases with Parallel Processing is industry data miners and practitioners in general, who would like to apply intelligent data mining techniques to large amounts of data. The book will also be of interest to academic researchers and postgraduate students, particularly database researchers, interested in advanced, intelligent database applications, and artificial intelligence researchers interested in industrial, real-world applications of machine learning.
Multimedia data comprising of images, audio and video is becoming increasingly common. The decreasing costs of consumer electronic devices such as digital cameras and digital camcorders, along with the ease of transportation facilitated by the Internet, has lead to a phenomenal rise in the amount of multimedia data generated and distributed. Given that this trend of increased use of multimedia data is likely to accelerate, there is an urgent need for providing a clear means of capturing, storing, indexing, retrieving, analyzing and summarizing such data. Content-based access to multimedia data is of primary importance since it is the natural way by which human beings interact with such information. To facilitate the content-based access of multimedia information, the first step is to derive feature measures from these data so that a feature space representation of the data content can be formed. This can subsequently allow for mapping the feature space to the symbol space (semantics) either automatically or through human intervention. Thus, signal to symbol mapping, useful for any practical system, can be successfully achieved. Perspectives on Content-Based Multimedia Systems provides a comprehensive set of techniques to tackle these important issues. This book offers detailed solutions to a wide range of practical problems in building real systems by providing specifics of three systems built by the authors. While providing a systems focus, it also equips the reader with a keen understanding of the fundamental issues, including a formalism for content-based multimedia database systems, multimedia feature extraction, object-based techniques, signature-based techniques and fuzzy retrieval techniques. The performance evaluation issues of practical systems is also explained. This book brings together essential elements of building a content-based multimedia database system in a way that makes them accessible to practitioners in computer science and electrical engineering. It can also serve as a textbook for graduate-level courses.
Database and Mobile Computing brings together in one place important contributions and up-to-date research results in this important area. Databases and Mobile Computing serves as an excellent reference, providing insight into some of the most important research issues in the field.
This book outlines the consequences of digitization for peer-reviewed research articles published in electronic journals. It has often been argued that digitization will revolutionize scientific communication. However, this study shows that this is not the case as far as scientific journals are concerned. Authors make little or no use of the possibilities offered by the digital medium, new procedures for electronic peer review have not replaced traditional peer review, and users do not seem to accept new forms of interaction offered by some electronic journals. The main innovations are to be found at the level of the infrastructures developed by publishers. Scientists themselves appear to be reluctant to change their established patterns of behaviour in formal scientific communication.
Spatial trajectories have been bringing the unprecedented wealth to a variety of research communities. A spatial trajectory records the paths of a variety of moving objects, such as people who log their travel routes with GPS trajectories. The field of moving objects related research has become extremely active within the last few years, especially with all major database and data mining conferences and journals. "Computing with Spatial Trajectories" introduces the algorithms, technologies, and systems used to process, manage and understand existing spatial trajectories for different applications. This book also presents an overview on both fundamentals and the state-of-the-art research inspired by spatial trajectory data, as well as a special focus on trajectory pattern mining, spatio-temporal data mining and location-based social networks. Each chapter provides readers with a tutorial-style introduction to one important aspect of location trajectory computing, case studies and many valuable references to other relevant research work. "Computing with Spatial Trajectories" is designed as a reference or secondary text book for advanced-level students and researchers mainly focused on computer science and geography. Professionals working on spatial trajectory computing will also find this book very useful.
The importance of having ef cient and effective methods for data mining and kn- ledge discovery (DM&KD), to which the present book is devoted, grows every day and numerous such methods have been developed in recent decades. There exists a great variety of different settings for the main problem studied by data mining and knowledge discovery, and it seems that a very popular one is formulated in terms of binary attributes. In this setting, states of nature of the application area under consideration are described by Boolean vectors de ned on some attributes. That is, by data points de ned in the Boolean space of the attributes. It is postulated that there exists a partition of this space into two classes, which should be inferred as patterns on the attributes when only several data points are known, the so-called positive and negative training examples. The main problem in DM&KD is de ned as nding rules for recognizing (cl- sifying) new data points of unknown class, i. e. , deciding which of them are positive and which are negative. In other words, to infer the binary value of one more attribute, called the goal or class attribute. To solve this problem, some methods have been suggested which construct a Boolean function separating the two given sets of positive and negative training data points.
Knowledge-based (KB) technology is being applied to complex problem-solving and critical tasks in many application domains. Concerns have naturally arisen as to the dependability of knowledge-based systems (KBS). As with any software, attention to quality and safety must be paid throughout development of a KBS and rigorous verification and validation (V&V) techniques must be employed. Research in V&V of KBS has emerged as a distinct field only in the last decade and is intended to address issues associated with quality and safety aspects of KBS and to credit such applications with the same degree of dependability as conventional applications. In recent years, V&V of KBS has been the topic of annual workshops associated with the main AI conferences, such as AAAI, IJACI and ECAI. Validation and Verification of Knowledge Based Systems contains a collection of papers, dealing with all aspects of KBS V&V, presented at the Fifth European Symposium on Verification and Validation of Knowledge Based Systems and Components (EUROVAV'99 - http: //www.dnv.no/research/safekbs/eurovav99/) which was held in Oslo in the summer of 1999, and was sponsored by Det Norske Veritas and the British Computer Society's Specialist Group on Expert Systems (SGES).
High Performance Data Mining: Scaling Algorithms, Applications and Systems brings together in one place important contributions and up-to-date research results in this fast moving area. High Performance Data Mining: Scaling Algorithms, Applications and Systems serves as an excellent reference, providing insight into some of the most challenging research issues in the field.
This is a compilation of papers presented at the Information System Concepts conference in Marburg, Germany. The special focus is consolidation and harmonisation of the numerous and widely diverging views in the field of information systems. This issue has become a hot topic, as many leading information system researchers and practitioners come to realise the importance of better communication among the members of the information systems community, and of a better scientific foundation of this rapidly evolving field.
This book provides a new direction in the field of nano-optics and nanophotonics from information and computing-related sciences and technology. Entitled by "Information Physics and Computing in NanosScale Photonics and Materials", IPCN in short, the book aims to bring together recent progresses in the intersection of nano-scale photonics, information, and enabling technologies. The topic will include (1) an overview of information physics in nanophotonics, (2) DNA self-assembled nanophotonic systems, (3) Functional molecular sensing, (4) Smart fold computing, an architecture for nanophotonics, (5) semiconductor nanowire and its photonic applications, (6) single photoelectron manipulation in imaging sensors, (6) hierarchical nanophotonic systems, (8) photonic neuromorphic computing, and (9) SAT solver and decision making based on nanophotonics.
Continuous improvements in data analysis and cloud computing have allowed more opportunities to develop systems with user-focused designs. This not only leads to higher success in day-to-day usage, but it increases the overall probability of technology adoption. Advancing Cloud Database Systems and Capacity Planning with Dynamic Applications is a key resource on the latest innovations in cloud database systems and their impact on the daily lives of people in modern society. Highlighting multidisciplinary studies on information storage and retrieval, big data architectures, and artificial intelligence, this publication is an ideal reference source for academicians, researchers, scientists, advanced level students, technology developers and IT officials. |
You may like...
BTEC Nationals Information Technology…
Jenny Phillips, Alan Jarvis, …
Paperback
R1,018
Discovery Miles 10 180
Applied Big Data Analytics and Its Role…
Peng Zhao, Xin Wang, …
Hardcover
R6,648
Discovery Miles 66 480
Management Of Information Security
Michael Whitman, Herbert Mattord
Paperback
Data Analytics for Social Microblogging…
Soumi Dutta, Asit Kumar Das, …
Paperback
R3,335
Discovery Miles 33 350
Database Principles - Fundamentals of…
Carlos Coronel, Keeley Crockett, …
Paperback
|