![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Databases
Great advances have been made in the database field. Relational and object- oriented databases, distributed and client/server databases, and large-scale data warehousing are among the more notable. However, none of these advances promises to have as great and direct an effect on the daily lives of ordinary citizens as video databases. Video databases will provide a quantum jump in our ability to deal with visual data, and in allowing people to access and manipulate visual information in ways hitherto thought impossible. Video Database Systems: Issues, Products and Applications gives practical information on academic research issues, commercial products that have already been developed, and the applications of the future driving this research and development. This book can also be considered a reference text for those entering the field of video or multimedia databases, as well as a reference for practitioners who want to identify the kinds of products needed in order to utilize video databases. Video Database Systems: Issues, Products and Applications covers concepts, products and applications. It is written at a level which is less detailed than that normally found in textbooks but more in-depth than that normally written in trade press or professional reference books. Thus, it seeks to serve both an academic and industrial audience by providing a single source of information about the research issues in the field, and the state-of-the-art of practice.
This book explores community dynamics within social media. Using Wikipedia as an example, the volume explores communities that rely upon commons-based peer production. Fundamental theoretical principles spanning such domains as organizational configurations, leadership roles, and social evolutionary theory are developed. In the context of Wikipedia, these theories explain how a functional elite of highly productive editors has emerged and why they are responsible for a majority of the content. It explains how the elite shapes the project and how this group tends to become stable and increasingly influential over time. Wikipedia has developed a new and resilient social hierarchy, an adhocracy, which combines features of traditional and new, online, social organizations. The book presents a set of practical approaches for using these theories in real-world practice. This work fundamentally changes the way we think about social media leadership and evolution, emphasizing the crucial contributions of leadership, of elite social roles, and of group global structure to the overall success and stability of large social media projects. Written in an accessible and direct style, the book will be of interest to academics as well as professionals with an interest in social media and commons-based peer production processes.
Steganography, a means by which two or more parties may communicate using "invisible" or "subliminal" communication, and watermarking, a means of hiding copyright data in images, are becoming necessary components of commercial multimedia applications that are subject to illegal use. This is a comprehensive survey of steganography and watermarking and their application to modern communications and multimedia. It helps the reader to understand steganography, the history of this previously neglected element of cryptography, the hurdles of international law on strong cryptographic techniques, and a description of the methods you can use to hide information in modern media. Included in this discussion is an overview of "steganalysis", methods which can be used to break stenographic communication. This resource also includes an introduction to and survey of watermarking methods, and discusses this method's similarities to and differences from steganography. The reader should gain a working knowledge of watermarking's pros and cons, and learn the legal implications of watermarking and copyright issues on the Internet.
In our increasingly mobile world the ability to access information on demand at any time and place can satisfy people's information needs as well as confer on them a competitive advantage. The emergence of battery-operated, low-cost and portable computers such as palmtops and PDAs, coupled with the availability and exploitation of wireless networks, have made possible the potential for ubiquitous computing. Through the wireless networks, portable equipments will become an integrated part of existing distributed computing environments, and mobile users can have access to data stored at information servers located at the static portion of the network even while they are on the move. Traditionally, information is retrieved following a request-response model. However, this model is no longer adequate in a wireless computing environment. First, the wireless channel is unreliable and the bandwidth is low compared to the wired counterpart. Second, the environment is essentially asymmetric with a large number of mobile users accessing a small number of servers. Third, battery-operated portable devices can typically operate only for a short time because of the short battery lifespan. Thus, clients are expected to be disconnected most of the time. To overcome these limitations, there has been a proliferation of research efforts on designing data delivery mechanisms to support wireless computing more effectively. Data Dissemination in Wireless Computing Environments focuses on such mechanisms. The purpose is to provide a thorough and comprehensive review of recent advances on energy-efficient data delivery protocols, efficient wireless channel bandwidth utilization, reliable broadcasting and cache invalidation strategies for clients with long disconnection time. Besides surveying existing methods, this book also compares and evaluates some of the more promising schemes.
This book presents the cyber culture of micro, macro, cosmological, and virtual computing. The book shows how these work to formulate, explain, and predict the current processes and phenomena monitoring and controlling technology in the physical and virtual space.The authors posit a basic proposal to transform description of the function truth table and structure adjacency matrix to a qubit vector that focuses on memory-driven computing based on logic parallel operations performance. The authors offer a metric for the measurement of processes and phenomena in a cyberspace, and also the architecture of logic associative computing for decision-making and big data analysis.The book outlines an innovative theory and practice of design, test, simulation, and diagnosis of digital systems based on the use of a qubit coverage-vector to describe the functional components and structures. Authors provide a description of the technology for SoC HDL-model diagnosis, based on Test Assertion Blocks Activated Graph. Examples of cyber-physical systems for digital monitoring and cloud management of social objects and transport are proposed. A presented automaton model of cosmological computing explains the cyclical and harmonious evolution of matter-energy essence, and also a space-time form of the Universe.
Many real-time systems rely on static scheduling algorithms. This includes cyclic scheduling, rate monotonic scheduling and fixed schedules created by off-line scheduling techniques such as dynamic programming, heuristic search, and simulated annealing. However, for many real-time systems, static scheduling algorithms are quite restrictive and inflexible. For example, highly automated agile manufacturing, command, control and communications, and distributed real-time multimedia applications all operate over long lifetimes and in highly non-deterministic environments. Dynamic real-time scheduling algorithms are more appropriate for these systems and are used in such systems. Many of these algorithms are based on earliest deadline first (EDF) policies. There exists a wealth of literature on EDF-based scheduling with many extensions to deal with sophisticated issues such as precedence constraints, resource requirements, system overload, multi-processors, and distributed systems. Deadline Scheduling for Real-Time Systems: EDF and Related Algorithms aims at collecting a significant body of knowledge on EDF scheduling for real-time systems, but it does not try to be all-inclusive (the literature is too extensive). The book primarily presents the algorithms and associated analysis, but guidelines, rules, and implementation considerations are also discussed, especially for the more complicated situations where mathematical analysis is difficult. In general, it is very difficult to codify and taxonomize scheduling knowledge because there are many performance metrics, task characteristics, and system configurations. Also, adding to the complexity is the fact that a variety of algorithms have beendesigned for different combinations of these considerations. In spite of the recent advances there are still gaps in the solution space and there is a need to integrate the available solutions. For example, a list of issues to consider includes: preemptive versus non-preemptive tasks, uni-processors versus multi-processors, using EDF at dispatch time versus EDF-based planning, precedence constraints among tasks, resource constraints, periodic versus aperiodic versus sporadic tasks, scheduling during overload, fault tolerance requirements, and providing guarantees and levels of guarantees (meeting quality of service requirements). Deadline Scheduling for Real-Time Systems: EDF and Related Algorithms should be of interest to researchers, real-time system designers, and instructors and students, either as a focussed course on deadline-based scheduling for real-time systems, or, more likely, as part of a more general course on real-time computing. The book serves as an invaluable reference in this fast-moving field.
This book presents a state-of-the art review of current perspectives on Communications and Multimedia Security. It contains the Proceedings of the 3rd Joint Working Conference of IFIP TC6 and TC11, arranged by the International Federation for Information Processing and held in Athens, Greece in September 1997. The book aims to cover the subject of Communications and Multimedia Systems Security, as fully as possible. It constitutes an essential reading for information technology security specialists; computer professionals; communication systems professionals; EDP managers; EDP auditors; managers, researchers and students working on the subject.
This book presents the combined peer-reviewed proceedings of the tenth International Symposium on Intelligent Distributed Computing (IDC'2016), which was held in Paris, France from October 10th to 12th, 2016. The 23 contributions address a range of topics related to theory and application of intelligent distributed computing, including: Intelligent Distributed Agent-Based Systems, Ambient Intelligence and Social Networks, Computational Sustainability, Intelligent Distributed Knowledge Representation and Processing, Smart Networks, Networked Intelligence and Intelligent Distributed Applications, amongst others.
Chaos-based cryptography, attracting many researchers in the past decade, is a research field across two fields, i.e., chaos (nonlinear dynamic system) and cryptography (computer and data security). It Chaos' properties, such as randomness and ergodicity, have been proved to be suitable for designing the means for data protection. The book gives a thorough description of chaos-based cryptography, which consists of chaos basic theory, chaos properties suitable for cryptography, chaos-based cryptographic techniques, and various secure applications based on chaos. Additionally, it covers both the latest research results and some open issues or hot topics. The book creates a collection of high-quality chapters contributed by leading experts in the related fields. It embraces a wide variety of aspects of the related subject areas and provide a scientifically and scholarly sound treatment of state-of-the-art techniques to students, researchers, academics, personnel of law enforcement and IT practitioners who are interested or involved in the study, research, use, design and development of techniques related to chaos-based cryptography.
Requirements engineering has since long acknowledged the importance of the notion that system requirements are stakeholder goals-rather than system functions-and ought to be elicited, modeled and analyzed accordingly. In this book, Nurcan and her co-editors collected twenty contributions from leading researchers in requirements engineering with the intention to comprehensively present an overview of the different perspectives that exist today, in 2010, on the concept of intention in the information systems community. These original papers honor Colette Rolland for her contributions to this field, as she was probably the first to emphasize that 'intention' has to be considered as a first-class concept in information systems engineering. Written by long-term collaborators (and most often friends) of Colette Rolland, this volume covers topics like goal-oriented requirements engineering, model-driven development, method engineering, and enterprise modeling. As such, it is a tour d'horizon of Colette Rolland's lifework, and is presented to her on the occasion of her retirement at CaISE 2010 in Hammamet, the conference she once cofounded and which she helped to grow and prosper for more than 20 years.
Current research in Visual Database Systems can be characterized by scalability, multi-modality of interaction, and higher semantic levels of data. Visual interfaces that allow users to interact with large databases must scale to web and distributed applications. Interaction with databases must employ multiple and more diversified interaction modalities, such as speech and gesture, in addition to visual exploitation. Finally, the basic elements managed in modern databases are rapidly evolving, from text, images, sound, and video, to compositions and now annotations of these media, thus incorporating ever-higher levels and different facets of semantics. In addition to visual interfaces and multimedia databases, Visual and Multimedia Information Management includes research in the following areas: Speech and aural interfaces to databases; Visualization of web applications and database structure; Annotation and retrieval of image databases; Visual querying in geographical information systems; Video databases; and Virtual environment and modeling of complex shapes. Visual and Multimedia Information Management comprises the proceedings of the sixth International Conference on Visual Database Systems, which was sponsored by the International Federation for Information Processing (IFIP), and held in Brisbane, Australia, in May 2002. This volume will be essential for researchers in the field of management of visual and multimedia information, as well as for industrial practitioners concerned with building IT products for managing visual and multimedia information.
From environmental management to land planning and geo-marketing, the number of application domains that may greatly benefit from using data enriched with spatio-temporal features is expanding very rapidly. Unfortunately, development of new spatio-temporal applications is hampered by the lack of conceptual design methods suited to cope with the additional complexity of spatio-temporal data. This complexity is obviously due to the particular semantics of space and time, but also to the need for multiple representations of the same reality to address the diversity of requirements from highly heterogeneous user communities. Conceptual design methods are also needed to facilitate the exchange and reuse of existing data sets, a must in geographical data management due to the high collection costs of the data. Yet, current practice in areas like geographical information systems or moving objects databases does not include conceptual design methods very well, if at all. This book shows that a conceptual design approach for spatio-temporal databases is both feasible and easy to apprehend. While providing a firm basis through extensive discussion of traditional data modeling concepts, the major focus of the book is on modeling spatial and temporal information. Parent, Spaccapietra and Zimanyi provide a detailed and comprehensive description of an approach that fills the gap between application conceptual requirements and system capabilities, covering both data modeling and data manipulation features. The ideas presented summarize several years of research on the characteristics and description of space, time, and perception. In addition to the authors' own data modeling approach, MADS (Modeling of Application Data with Spatio-temporal features), the book also surveys alternative data models and approaches (from industry and academia) that target support of spatio-temporal modeling. The reader will acquire intimate knowledge of both the traditional and innovative features that form a consistent data modeling approach. Visual notations and examples are employed extensively to illustrate the use of the various constructs. Therefore, this book is of major importance and interest to advanced professionals, researchers, and graduate or post-graduate students in the areas of spatio-temporal databases and geographical information systems. "For anyone thinking of doing research in this field, or who is developing a system based on spatio-temporal data, this text is essential reading." (Mike Worboys, U Maine, Orono, ME, USA) "The high-level semantic model presented and validated in this book provides essential guidance to researchers and implementers when improving the capabilities of data systems to serve the actual needs of applications and their users in the temporal and spatial domains that are so prevalent today." (Gio Wiederhold, Stanford U, CA, USA)"
Welcome to the 6th International Conference on Open Source Systems of the IFIP Working Group 2. 13. This year was the ?rst time this international conf- ence was held in North America. We had a large number of high-quality papers, highlyrelevantpanelsandworkshops, acontinuationofthepopulardoctoralc- sortium, and multiple distinguished invited speakers. The success of OSS 2010 was only possible because an Organizing Committee, a Program Committee, Workshop and Doctoral Committees, and authors of research manuscripts from over 25 countries contributed their time and interest to OSS 2010. In the spirit of the communities we study, you self-organized, volunteered, and contributed to this important research forum studying free, libre, open source software and systems. We thank you Despite our modest success, we have room to improve and grow our conf- ence and community. At OSS 2010 we saw little or no participation from large portions of the world, including Latin America, Africa, China, and India. But opportunitiestoexpandarepossible. InJapan, weseeahotspotofparticipation led by Tetsuo Noda and his colleagues, both with full-paper submissions and a workshopon"OpenSourcePolicyandPromotionofITIndustries inEastAsia. " The location of OSS 2011 in Salvador, Brazil, will hopefully result in signi?cant participation from researchers in Brazil - already a strong user of OSS - and otherSouthAmericancountries. UndertheleadershipofMeganSquire, Publicity Chair, we recruited RegionalPublicity Co-chairscovering Japan (Tetsuo Noda), Africa(SulaymanSowe), the MiddleEastandSouthAsia(FaheenAhmed), R- sia and Eastern Europe (Alexey Khoroshilov), Western Europe (Yeliz Eseryel), UK and Ireland (Andrea Capiluppi), and the Nordic countries (Bj] orn Lundell)."
The present book outlines a new approach to possibilistic clustering in which the sought clustering structure of the set of objects is based directly on the formal definition of fuzzy cluster and the possibilistic memberships are determined directly from the values of the pairwise similarity of objects. The proposed approach can be used for solving different classification problems. Here, some techniques that might be useful at this purpose are outlined, including a methodology for constructing a set of labeled objects for a semi-supervised clustering algorithm, a methodology for reducing analyzed attribute space dimensionality and a methods for asymmetric data processing. Moreover, a technique for constructing a subset of the most appropriate alternatives for a set of weak fuzzy preference relations, which are defined on a universe of alternatives, is described in detail, and a method for rapidly prototyping the Mamdani s fuzzy inference systems is introduced. This book addresses engineers, scientists, professors, students and post-graduate students, who are interested in and work with fuzzy clustering and its applications
During the past few years, data mining has grown rapidly in visibility and importance within information processing and decision analysis. This is par ticularly true in the realm of e-commerce, where data mining is moving from a "nice-to-have" to a "must-have" status. In a different though related context, a new computing methodology called granular computing is emerging as a powerful tool for the conception, analysis and design of information/intelligent systems. In essence, data mining deals with summarization of information which is resident in large data sets, while granular computing plays a key role in the summarization process by draw ing together points (objects) which are related through similarity, proximity or functionality. In this perspective, granular computing has a position of centrality in data mining. Another methodology which has high relevance to data mining and plays a central role in this volume is that of rough set theory. Basically, rough set theory may be viewed as a branch of granular computing. However, its applications to data mining have predated that of granular computing."
The most important use of computing in the future will be in the context of the global "digital convergence" where everything becomes digital and every thing is inter-networked. The application will be dominated by storage, search, retrieval, analysis, exchange and updating of information in a wide variety of forms. Heavy demands will be placed on systems by many simultaneous re quests. And, fundamentally, all this shall be delivered at much higher levels of dependability, integrity and security. Increasingly, large parallel computing systems and networks are providing unique challenges to industry and academia in dependable computing, espe cially because of the higher failure rates intrinsic to these systems. The chal lenge in the last part of this decade is to build a systems that is both inexpensive and highly available. A machine cluster built of commodity hardware parts, with each node run ning an OS instance and a set of applications extended to be fault resilient can satisfy the new stringent high-availability requirements. The focus of this book is to present recent techniques and methods for im plementing fault-tolerant parallel and distributed computing systems. Section I, Fault-Tolerant Protocols, considers basic techniques for achieving fault-tolerance in communication protocols for distributed systems, including synchronous and asynchronous group communication, static total causal order ing protocols, and fail-aware datagram service that supports communications by time."
Multimedia is changing the design of database and information retrieval systems. The accumulation of audio, image, and video content is of little use in these systems if the content cannot be retrieved on demand, a critical requirement that has led to the development of new technologies for the analysis and indexing of media data. In turn, these technologies seek to derive information or features from a data type that can facilitate rapid retrieval, efficient compression, and logical presentation of the data. Significant work that has not been addressed, however, is the benefits of analyzing more than one data type simultaneously. Computed Synchronization for Multimedia Applications presents a new framework for the simultaneous analysis of multiple media data objects. The primary benefit of this analysis is computed synchronization, a temporal and spatial alignment of multiple media objects. Computed Synchronization for Multimedia Applications also presents several specific applications and a general structure for the solution of computed synchronization problems. The applications demonstrate the use of this structure. Two applications in particular are described in detail: the alignment of text to speech audio, and the alignment of simultaneous English language translations of ancient texts. Many additional applications are discussed as future uses of the technology. Computed Synchronization for Multimedia Applications is useful to researchers, students, and developers seeking to apply computed synchronization in many fields. It is also suitable as a reference for a graduate-level course in multimedia data retrieval.
There is a growing interest in integrating databases and programming languages. In recent years the programming language community has developed new models of computation such as logic programming, object-oriented programming and functional programming, to add to the well established von Neumann model. The data base community has almost independently developed more and more sophisticated data models to solve the problems of large scale data organisation. To make use of these new models in programming languages there must be an awareness of the problems of large scale data. The data base designers can also learn much about language interfaces from programming language designers. The purpose of this book is to present the state of the art in integrating both approaches. The book evolved from the proceedings of a workshop held at the Appin in August 1985. It consists of three sections. The first, "Data Types and Persistence," discusses the issues of data abstraction in a persistent environment. Type systems, modules and binding mechanisms that are appropriate for programming in the large are proposed. Type checking for polymorphic systems and across innovations of the type checker are also discussed. The second section, "Database Types in Programming Languages," introduces the concept of inheritance as a method of polymorphic modelling. It is shown how inheritance can be used as a method of computation in logic programming and how it is appropriate for modelling large scale data in databases. The last section discusses the issues of controlled access to large scale data in a concurrent and distributed persistent environment. Finally methods of how we may implement persistence and buildmachine architectures for persistent data round off the book.
Business rules are everywhere. Every enterprise process, task, activity, or function is governed by rules. However, some of these rules are implicit and thus poorly enforced, others are written but not enforced, and still others are perhaps poorly written and obscurely enforced. The business rule approach looks for ways to elicit, communicate, and manage business rules in a way that all stakeholders can understand, and to enforce them within the IT infrastructure in a way that supports their traceability and facilitates their maintenance. Boyer and Mili will help you to adopt the business rules approach effectively. While most business rule development methodologies put a heavy emphasis on up-front business modeling and analysis, agile business rule development (ABRD) as introduced in this book is incremental, iterative, and test-driven. Rather than spending weeks discovering and analyzing rules for a complete business function, ABRD puts the emphasis on producing executable, tested rule sets early in the project without jeopardizing the quality, longevity, and maintainability of the end result. The authors presentation covers all four aspects required for a successful application of the business rules approach: (1) foundations, to understand what business rules are (and are not) and what they can do for you; (2) methodology, to understand how to apply the business rules approach; (3) architecture, to understand how rule automation impacts your application; (4) implementation, to actually deliver the technical solution within the context of a particular business rule management system (BRMS). Throughout the book, the authors use an insurance case study that deals with claim processing. Boyer and Mili cater to different audiences: Project managers will find a pragmatic, proven methodology for delivering and maintaining business rule applications. Business analysts and rule authors will benefit from guidelines and best practices for rule discovery and analysis. Application architects and software developers will appreciate an exploration of the design space for business rule applications, proven architectural and design patterns, and coding guidelines for using JRules.
This book reports on advanced theories and cutting-edge applications in the field of soft computing. The individual chapters, written by leading researchers, are based on contributions presented during the 4th World Conference on Soft Computing, held May 25-27, 2014, in Berkeley. The book covers a wealth of key topics in soft computing, focusing on both fundamental aspects and applications. The former include fuzzy mathematics, type-2 fuzzy sets, evolutionary-based optimization, aggregation and neural networks, while the latter include soft computing in data analysis, image processing, decision-making, classification, series prediction, economics, control, and modeling. By providing readers with a timely, authoritative view on the field, and by discussing thought-provoking developments and challenges, the book will foster new research directions in the diverse areas of soft computing.
This book explores the concepts of data mining and data warehousing, a promising and flourishing frontier in database systems, and presents a broad, yet in-depth overview of the field of data mining. Data mining is a multidisciplinary field, drawing work from areas including database technology, artificial intelligence, machine learning, neural networks, statistics, pattern recognition, knowledge based systems, knowledge acquisition, information retrieval, high performance computing and data visualization.
Multimedia Cartography provides a contemporary overview of the issues related to multimedia cartography and the design and production elements that are unique to this area of mapping. The book has been written for professional cartographers interested in moving into multimedia mapping, for cartographers already involved in producing multimedia titles who wish to discover the approaches that other practitioners in multimedia cartography have taken and for students and academics in the mapping sciences and related geographical fields wishing to update their knowledge about current issues related to cartographic design and production. It provides a new approach to cartography one based on the exploitation of the many rich media components and avant-garde approach that multimedia offers."
Universal navigation is accessible primarily through smart phones providing users with navigation information regardless of the environment (i.e., outdoor or indoor). Universal Navigation for Smart Phones provide the most up-to-date navigation technologies and systems for both outdoor and indoor navigation. It also provides a comparison of the similarities and differences between outdoor and indoor navigation systems from both a technological stand point and user 's perspective. All aspects of navigation systems including geo-positioning, wireless communication, databases, and functions will be introduced. The main thrust of this book presents new approaches and techniques for future navigation systems including social networking, as an emerging approach for navigation.
A collection of the most up-to-date research-oriented chapters on information systems development and database, this book provides an understanding of the capabilities and features of new ideas and concepts in information systems development, databases, and forthcoming technologies. |
You may like...
Mathematical Foundations of Advanced…
Bernhard Steffen, Oliver Ruthing, …
Hardcover
R1,387
Discovery Miles 13 870
Creativity in Computing and DataFlow…
Suyel Namasudra, Veljko Milutinovic
Hardcover
R4,204
Discovery Miles 42 040
Advances in Time Series Analysis and…
Ignacio Rojas, Hector Pomares, …
Hardcover
R2,718
Discovery Miles 27 180
Continued Fractions with Applications…
L. Lorentzen, H. Waadeland
Hardcover
R1,386
Discovery Miles 13 860
Algebraic Methods in General Rough Sets
aMani, Gianpiero Cattaneo, …
Hardcover
R3,483
Discovery Miles 34 830
Advanced Topics in Bisimulation and…
Davide Sangiorgi, Jan Rutten
Hardcover
R3,205
Discovery Miles 32 050
Parallelism in Matrix Computations
Efstratios Gallopoulos, Bernard Philippe, …
Hardcover
R3,621
Discovery Miles 36 210
|