![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming
This book describes a novel methodology for studying algorithmic skills, intended as cognitive activities related to rule-based symbolic transformation, and argues that some human computational abilities may be interpreted and analyzed as genuine examples of extended cognition. It shows that the performance of these abilities relies not only on innate neurocognitive systems or language-related skills, but also on external tools and general agent-environment interactions. Further, it asserts that a low-level analysis, based on a set of core neurocognitive systems linking numbers and language, is not sufficient to explain some specific forms of high-level numerical skills, like those involved in algorithm execution. To this end, it reports on the design of a cognitive architecture for modeling all the relevant features involved in the execution of algorithmic strategies, including external tools, such as paper and pencils. The first part of the book discusses the philosophical premises for endorsing and justifying a position in philosophy of mind that links a modified form of computationalism with some recent theoretical and scientific developments, like those introduced by the so-called dynamical approach to cognition. The second part is dedicated to the description of a Turing-machine-inspired cognitive architecture, expressly designed to formalize all kinds of algorithmic strategies.
This book constitutes the thoroughly refereed post-conference proceedings of the 11th IFIP WG 6.11 Conference on e-Business, e-Services and e-Society, I3E 2011, held in Kaunas, Lithuania, in October 2011. The 25 revised papers presented were carefully reviewed and selected from numerous submissions. They are organized in the following topical sections: e-government and e-governance, e-services, digital goods and products, e-business process modeling and re-engineering, innovative e-business models and implementation, e-health and e-education, and innovative e-business models.
Research on real-time Java technology has been prolific over the past decade, leading to a large number of corresponding hardware and software solutions, and frameworks for distributed and embedded real-time Java systems. This book is aimed primarily at researchers in real-time embedded systems, particularly those who wish to understand the current state of the art in using Java in this domain. Much of the work in real-time distributed, embedded and real-time Java has focused on the Real-time Specification for Java (RTSJ) as the underlying base technology, and consequently many of the Chapters in this book address issues with, or solve problems using, this framework.Describes innovative techniques in: scheduling, memory management, quality of service and communication systems supporting real-time Java applications;Includes coverage of multiprocessor embedded systems and parallel programming;Discusses state-of-the-art resource management for embedded systems, including Java's real-time garbage collection and parallel collectors;Considers hardware support for the execution of Java programs including how programs can interact with functional accelerators;Includes coverage of Safety Critical Java for development of safety critical embedded systems."
Process calculi are among the most successful models of concurrent systems. Various behavior equivalences between processes are central notions in CCS (calculus of communicating systems) and other process calculi. In the real applications, specification and implementation are described as two processes, and correctness of programs is treated as a certain behavior equivalence between them. The purpose of this book is to establish a theory of approximate correctness and infinite evolution of concurrent programs by employing some notions and tools from point-set topology. This book is restricted to CCS for simplicity, but the main idea also applies to some other process calculi. The concept of bisimulation limits, useful for the understanding and analysis of infinite evolution of processes, is introduced. In addition, the notions of near bisimulations and bisimulation indexes, suitable in describing approximate correctness of concurrent programs, are proposed. The book will be of particular interest to researchers in the fields of theoretical computer science, especially theory of concurrency and hybrid systems, and graduate students in related disciplines. It will also be valuable to practical system designers developing concurrent and/or real-time systems.
Despite differing origins, constraint programming and mathematical
programming are beginning to merge. Constraint programming has
grown out of the logic programming community as part of an effort
to embed constraints in a programming language. Mathematical
programming, a much older field, is rooted in the mathematics of
optimization. Because these two areas have complementary strengths,
there are ongoing efforts to integrate the two.Constraint and
Integer Programming presents some of the basic ideas of constraint
programming and mathematical programming, explores approaches to
integration, brings us up to date on heuristic methods, and
attempts to discern future directions in this fast-moving field.
This book presents the latest research findings and state-of-the-art solutions on optimization techniques and provides new research direction and developments. Both the theoretical and practical aspects of the book will be much beneficial to experts and students in optimization and operation research community. It selects high quality papers from The International Conference on Optimization: Techniques and Applications (ICOTA2013). The conference is an official conference series of POP (The Pacific Optimization Research Activity Group; there are over 500 active members). These state-of-the-art works in this book authored by recognized experts will make contributions to the development of optimization with its applications.
Soft City Culture and Technology: The Betaville Project discusses the complete cycle of conception, development, and deployment of the Betaville platform. Betaville is a massively participatory online environment for distributed 3D design and development of proposals for changes to the built environment an experimental integration of art, design, and software development for the public realm. Through a detailed account of Betaville from a Big Crazy Idea to a working "deep social medium," the author examines the current conditions of performance and accessibility of hardware, software, networks, and skills that can be brought together into a new form of open public design and deliberation space, for and spanning and integrating the disparate spheres of art, architecture, social media, and engineering. Betaville is an ambitious enterprise, of building compelling and constructive working relationships in situations where roles and disciplinary boundaries must be as agile as the development process of the software itself. Through a considered account and analysis of the interdependencies between Betaville's project design, development methods, and deployment, the reader can gain a deeper understanding of the potential socio-technical forms of New Soft Cities: blended virtual-physical worlds, whose "public works" must ultimately serve and succeed as massively collaborative works of art and infrastructure."
As the first book devoted to relational data mining, this coherently written multi-author monograph provides a thorough introduction and systematic overview of the area. The first part introduces the reader to the basics and principles of classical knowledge discovery in databases and inductive logic programming; subsequent chapters by leading experts assess the techniques in relational data mining in a principled and comprehensive way; finally, three chapters deal with advanced applications in various fields and refer the reader to resources for relational data mining.This book will become a valuable source of reference for R&D professionals active in relational data mining. Students as well as IT professionals and ambitioned practitioners interested in learning about relational data mining will appreciate the book as a useful text and gentle introduction to this exciting new field.
Cyberspace security is a critical subject of our times. On the one hand the development of Internet, mobile communications, distributed computing, computer software and databases storing essential enterprise information has helped to conduct business and personal communication between individual people. On the other hand it has created many opportunities for abuse, fraud and expensive damage. This book is a selection of the best papers presented at the NATO Advanced Research Workshop dealing with the Subject of Cyberspace Security and Defense. The level of the individual contributions in the volume is advanced and suitable for senior and graduate students, researchers and technologists who wish to get some feeling of the state of the art in several sub-disciplines of Cyberspace security. Several papers provide a broad-brush description of national security issues and brief summaries of technology states. These papers can be read and appreciated by technically enlightened managers and executives who want to understand security issues and approaches to technical solutions. An important question of our times is not "Should we do something for enhancing our digital assets security," the question is "How to do it."
The design process of embedded systems has changed substantially in recent years. One of the main reasons for this change is the pressure to shorten time-to-market when designing digital systems. To shorten the product cycles, programmable processes are used to implement more and more functionality of the embedded system. Therefore, nowadays, embedded systems are very often implemented by heterogeneous systems consisting of ASICs, processors, memories and peripherals. As a consequence, the research topic of hardware/software co-design, dealing with the problems of designing these heterogeneous systems, has gained great importance. Hardware/Software Co-design for Data Flow Dominated Embedded Systems introduces the different tasks of hardware/software co-design including system specification, hardware/software partitioning, co-synthesis and co-simulation. The book summarizes and classifies state-of-the-art co-design tools and methods for these tasks. In addition, the co-design tool COOL is presented which solves the co-design tasks for the class of data-flow dominated embedded systems. In Hardware/Software Co-design for Data Flow Dominated Embedded Systems the primary emphasis has been put on the hardware/software partitioning and the co-synthesis phase and their coupling. In contrast to many other publications in this area, a mathematical formulation of the hardware/software partitioning problem is given. This problem formulation supports target architectures consisting of multiple processors and multiple ASICs. Several novel approaches are presented and compared for solving the partitioning problem, including an MILP approach, a heuristic solution and an approach based on geneticalgorithms. The co-synthesis phase is based on the idea of controlling the system by means of a static run-time scheduler implemented in hardware. New algorithms are introduced which generate a complete set of hardware and software specifications required to implement heterogeneous systems. All of these techniques are described in detail and exemplified. Hardware/Software Co-design for Data Flow Dominated Embedded Systems is intended to serve students and researchers working on hardware/software co-design. At the same time the variety of presented techniques automating the design tasks of hardware/software systems will be of interest to industrial engineers and designers of digital systems. From the foreword by Peter Marwedel: Niemann's method should be known by all persons working in the field. Hence, I recommend this book for everyone who is interested in hardware/software co-design.
Semantic Web services promise to automate tasks such as discovery, mediation, selection, composition, and invocation of services, enabling fully flexible automated e-business. Their usage, however, still requires a significant amount of human intervention due to the lack of support for a machine-processable description. In this book, Jos de Bruijn and his coauthors lay the foundations for understanding the requirements that shape the description of the various aspects related to Semantic Web services, such as the static background knowledge in the form of ontologies, the functional description of the service, and the behavioral description of the service. They introduce the Web Service Modeling Language (WSML), which provides means for describing the functionality and behavior of Web services, as well as the underlying business knowledge, in the form of ontologies, with a conceptual grounding in the Web Service Modeling Ontology. Academic and industrial researchers as well as professionals will find a comprehensive overview of the concepts and challenges in the area of Semantic Web services, the Web Services Modeling Language and its relation to the Web Services Modeling Ontology, and an in-depth treatment of both enabling technologies and theoretical foundations.
PREVIOUS EDITIONThis textbook introduces the "Fundamentals of Multimedia", addressing real issues commonly faced in the workplace. The essential concepts are explained in a practical way to enable students to apply their existing skills to address problems in multimedia. Fully revised and updated, this new edition now includes coverage of such topics as 3D TV, social networks, high-efficiency video compression and conferencing, wireless and mobile networks, and their attendant technologies. Features: presents an overview of the key concepts in multimedia, including color science; reviews lossless and lossy compression methods for image, video and audio data; examines the demands placed by multimedia communications on wired and wireless networks; discusses the impact of social media and cloud computing on information sharing and on multimedia content search and retrieval; includes study exercises at the end of each chapter; provides supplementary resources for both students and instructors at an associated website.
The book provides a platform for dealing with the flaws and failings of the soft computing paradigm through different manifestations. The different chapters highlight the necessity of the hybrid soft computing methodology in general with emphasis on several application perspectives in particular. Typical examples include (a) Study of Economic Load Dispatch by Various Hybrid Optimization Techniques, (b) An Application of Color Magnetic Resonance Brain Image Segmentation by Para Optimus LG Activation Function, (c) Hybrid Rough-PSO Approach in Remote Sensing Imagery Analysis, (d) A Study and Analysis of Hybrid Intelligent Techniques for Breast Cancer Detection using Breast Thermograms, and (e) Hybridization of 2D-3D Images for Human Face Recognition. The elaborate findings of the chapters enhance the exhibition of the hybrid soft computing paradigm in the field of intelligent computing.
This text helps the reader generate clear, effective documentation that is tailored to the information requirements of the end-user. Written for technical writers and their managers, quality assurance experts, and software engineers, the book describes a user-centered information design method (UCID) that should help ensure documentation conveys significant information for the user. The UCID shows how to: integrate the four major information components of a software system - user interface labels, messages, online and printed documentation; make sure these elements work together to improve usability; deploy iterative design and prototyping procedures that minimize flaws and save time and money; and guide technical writers effectively.
This book bridges the widening gap between two crucial constituents of computational intelligence: the rapidly advancing technologies of machine learning in the digital information age, and the relatively slow-moving field of general-purpose search and optimization algorithms. With this in mind, the book serves to offer a data-driven view of optimization, through the framework of memetic computation (MC). The authors provide a summary of the complete timeline of research activities in MC - beginning with the initiation of memes as local search heuristics hybridized with evolutionary algorithms, to their modern interpretation as computationally encoded building blocks of problem-solving knowledge that can be learned from one task and adaptively transmitted to another. In the light of recent research advances, the authors emphasize the further development of MC as a simultaneous problem learning and optimization paradigm with the potential to showcase human-like problem-solving prowess; that is, by equipping optimization engines to acquire increasing levels of intelligence over time through embedded memes learned independently or via interactions. In other words, the adaptive utilization of available knowledge memes makes it possible for optimization engines to tailor custom search behaviors on the fly - thereby paving the way to general-purpose problem-solving ability (or artificial general intelligence). In this regard, the book explores some of the latest concepts from the optimization literature, including, the sequential transfer of knowledge across problems, multitasking, and large-scale (high dimensional) search, systematically discussing associated algorithmic developments that align with the general theme of memetics. The presented ideas are intended to be accessible to a wide audience of scientific researchers, engineers, students, and optimization practitioners who are familiar with the commonly used terminologies of evolutionary computation. A full appreciation of the mathematical formalizations and algorithmic contributions requires an elementary background in probability, statistics, and the concepts of machine learning. A prior knowledge of surrogate-assisted/Bayesian optimization techniques is useful, but not essential.
Over the past two decades, software engineering has come a long way from object-based to object-oriented to component-based design and development. Invasive software composition is a new technique that unifies and extends recent software engineering concepts like generic programming, aspect-oriented development, architecture systems, or subject-oriented development. To improve reuse, this new method regards software components as grayboxes and integrates them during composition. Building on a minimal set of program transformations, composition operator libraries can be developed that parameterize, extend, connect, mediate, and aspect-weave components. The book is centered around the JAVA language and the freely available demonstrator library COMPOST. It provides a wealth of materials for researchers, students, and professional software architects alike.
Recent growth in knowledge management concepts has played a vital role in the improvement of organizational performance. These knowledge management approaches have been influential in achieving the goal of efficient production of software development processes. Knowledge-Based Processes in Software Development focuses on the inherent issues to help practitioners in gaining understanding of software development processes. The best practices highlighted in this publication will be essential to software professionals working in the industry as well as students and researchers in the domain of software engineering in order to successfully employ knowledge management procedures.
The growing demand for high quality, safety, and security of software systems can only be met by rigorous application of formal methods during software design. Tools for formal methods in general, however, do not provide a sufficient level of automatic processing. This book methodically investigates the potential of first-order logic automated theorem provers for applications in software engineering.Illustrated by complete case studies on verification of communication and security protocols and logic-based component reuse, the book characterizes proof tasks to allow an assessment of the provers capabilities. Necessary techniques and extensions, e.g., for handling inductive and modal proof tasks, or for controlling the prover, are covered in detail. The book demonstrates that state-of-the-art automated theorem provers are capable of automatically handling important tasks during the development of high-quality software and it provides many helpful techniques for increasing practical usability of the automated theorem prover for successful applications.
This book provides a comprehensive and practically minded introduction into serious games for law enforcement agencies. Serious games offer wide ranging benefits for law enforcement with applications from professional trainings to command-level decision making to the preparation for crises events. This book explains the conceptual foundations of virtual and augmented reality, gamification and simulation. It further offers practical guidance on the process of serious games development from user requirements elicitation to evaluation. The chapters are intended to provide principles, as well as hands-on knowledge to plan, design, test and apply serious games successfully in a law enforcement environment. A diverse set of case studies showcases the enormous variety that is possible in serious game designs and application areas and offers insights into concrete design decisions, design processes, benefits and challenges. The book is meant for law enforcement professionals interested in commissioning their own serious games as well as game designers interested in collaborative pedagogy and serious games for the law enforcement and security sector.
The book offers a detailed guide to temporal ordering, exploring open problems in the field and providing solutions and extensive analysis. It addresses the challenge of automatically ordering events and times in text. Aided by TimeML, it also describes and presents concepts relating to time in easy-to-compute terms. Working out the order that events and times happen has proven difficult for computers, since the language used to discuss time can be vague and complex. Mapping out these concepts for a computational system, which does not have its own inherent idea of time, is, unsurprisingly, tough. Solving this problem enables powerful systems that can plan, reason about events, and construct stories of their own accord, as well as understand the complex narratives that humans express and comprehend so naturally. This book presents a theory and data-driven analysis of temporal ordering, leading to the identification of exactly what is difficult about the task. It then proposes and evaluates machine-learning solutions for the major difficulties. It is a valuable resource for those working in machine learning for natural language processing as well as anyone studying time in language, or involved in annotating the structure of time in documents.
Everything should be made as simple as possible, but not simpler. (Albert Einstein, Readers Digest, 1977) The modern practice of creating technical systems and technological processes of high effi.ciency besides the employment of new principles, new materials, new physical effects and other new solutions ( which is very traditional and plays the key role in the selection of the general structure of the object to be designed) also includes the choice of the best combination for the set of parameters (geometrical sizes, electrical and strength characteristics, etc.) concretizing this general structure, because the Variation of these parameters ( with the structure or linkage being already set defined) can essentially affect the objective performance indexes. The mathematical tools for choosing these best combinations are exactly what is this book about. With the advent of computers and the computer-aided design the pro bations of the selected variants are usually performed not for the real examples ( this may require some very expensive building of sample op tions and of the special installations to test them ), but by the analysis of the corresponding mathematical models. The sophistication of the mathematical models for the objects to be designed, which is the natu ral consequence of the raising complexity of these objects, greatly com plicates the objective performance analysis. Today, the main (and very often the only) available instrument for such an analysis is computer aided simulation of an object's behavior, based on numerical experiments with its mathematical model."
The implementation of object-oriented languages has been an active topic of research since the 1960s when the first Simula compiler was written. The topic received renewed interest in the early 1980s with the growing popularity of object-oriented programming languages such as c++ and Smalltalk, and got another boost with the advent of Java. Polymorphic calls are at the heart of object-oriented languages, and even the first implementation of Simula-67 contained their classic implementation via virtual function tables. In fact, virtual function tables predate even Simula-for example, Ivan Sutherland's Sketchpad drawing editor employed very similar structures in 1960. Similarly, during the 1970s and 1980s the implementers of Smalltalk systems spent considerable efforts on implementing polymorphic calls for this dynamically typed language where virtual function tables could not be used. Given this long history of research into the implementation of polymorphic calls, and the relatively mature standing it achieved over time, why, one might ask, should there be a new book in this field? The answer is simple. Both software and hardware have changed considerably in recent years, to the point where many assumptions underlying the original work in this field are no longer true. In particular, virtual function tables are no longer sufficient to implement polymorphic calls even for statically typed languages; for example, Java's interface calls cannot be implemented this way. Furthermore, today's processors are deeply pipelined and can execute instructions out-of order, making it difficult to predict the execution time of even simple code sequences."
Overtheyears, research inthelifescienceshasbenefitedgreatlyfromthequantita tive toolsofmathematics and modeling. Many aspectsofcomplex biological systems can be more deeply understood when mathematical techniques are incorporated into a scientific investigation. Modelingcanbefruitfully applied in many typesofbiological research, from studies on the molecular, cellular, and organ level, to experiments in wholeanimalsandinpopulations. Using the field of nutrition as an example, one can find many cases of recent advances in knowledge and understanding that were facilitated by the application of mathematical modelingtokineticdata. Theavailabilityofbiologicallyimportantstable isotope-labeled compounds, developments in sensitive mass spectrometry and other analytical techniques, and advances in the powerful modeling software applied to data haveeachcontributed toourability tocarryoutevermoresophisticated kinetic studies that are relevant to nutrition and the health sciences at many levels oforganization. Furthermore, weanticipatethatmodeling isonthebrinkofanothermajoradvance: the application of kinetic modeling to clinical practice. With advances in the abilityof modelstoaccesslargedatabases(e. g., apopulationofindividualpatientrecords)andthe developmentofuserinterfaces thatare"friendly"enough tobeused byclinicians who arenotmodelers, wepredictthathealthapplicationsmodeling willbeanimportantnew 51 directionformodelinginthe21 century. This book contains manuscripts that are based on presentations at the seventh conference in a series focused on advancing nutrition and health research by fostering exchange among scientists from such disciplines as nutrition, biology, mathematics, statistics, kinetics, andcomputing. Thethemesofthesixpreviousconferencesincluded general nutritionmodeling(CanoltyandCain, 1985;Hoover-PlowandChandra, 1988), amino acids and carbohydrates (Aburnrad, 1991), minerals (Siva Subramanian and Wastney, 1995), vitamins, proteins, andmodelingtheory(CoburnandTownsend, 1996), and physiological compartmental modeling (Clifford and Muller, 1998). The seventh conference in the series was held at The Pennsylvania State University from July 29 throughAugust1,2000. Themeetingbeganwithaninstructiveandentertainingkeynote address by Professor Britton Chance, Eldridge Reeves Johnson University Professor Emeritus of Biophysics, Physical Chemistry, and Radiologic Physics, University of Pennsylvania. Dr."
This 2nd edition has been completely revised and updated, with additional new chapters. It presents state-of-the-art research in this area and focuses on key topics such as: visualization of semantic and structural information and metadata in the context of the emerging Semantic Web; Ontology-based Information Visualization and the use of graphically represented ontologies; Semantic Visualizations using Topic Maps and graph techniques; Recommender systems for filtering and recommending on the Semantic Web; SVG and X3D as new XML-based languages for 2D and 3D visualisations; methods used to construct and visualize high quality metadata and ontologies; and navigating and exploring XML documents using interactive multimedia interfaces. The design of visual interfaces for e-commerce and information retrieval is currently a challenging area of practical web development.
Basics of Software Engineering Experimentation is a practical guide to experimentation in a field which has long been underpinned by suppositions, assumptions, speculations and beliefs. It demonstrates to software engineers how Experimental Design and Analysis can be used to validate their beliefs and ideas. The book does not assume its readers have an in-depth knowledge of mathematics, specifying the conceptual essence of the techniques to use in the design and analysis of experiments and keeping the mathematical calculations clear and simple. Basics of Software Engineering Experimentation is practically oriented and is specially written for software engineers, all the examples being based on real and fictitious software engineering experiments. |
You may like...
Principles of Big Graph: In-depth…
Ripon Patgiri, Ganesh Chandra Deka, …
Hardcover
R3,925
Discovery Miles 39 250
Clean Architecture - A Craftsman's Guide…
Robert Martin
Paperback
(1)
Essential Java for Scientists and…
Brian Hahn, Katherine Malan
Paperback
R1,266
Discovery Miles 12 660
Hardware Accelerator Systems for…
Shiho Kim, Ganesh Chandra Deka
Hardcover
R3,950
Discovery Miles 39 500
|