![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Applications of computing > Artificial intelligence > Knowledge-based systems / expert systems
The field of network programming is so large, and developing so rapidly, that it can appear almost overwhelming to those new to the discipline. Answering the need for an accessible overview of the field, this text/reference presents a manageable introduction to both the theoretical and practical aspects of computer networks and network programming. Clearly structured and easy to follow, the book describes cutting-edge developments in network architectures, communication protocols, and programming techniques and models, supported by code examples for hands-on practice with creating network-based applications. Topics and features: presents detailed coverage of network architectures, including the latest wireless heterogeneous networks, communication protocols, and support for communication-based services; gently introduces the reader to the basic ideas underpinning computer networking, before gradually building up to more advanced concepts; provides numerous step-by-step descriptions of practical examples in tandem with the theoretical discussions; examines a range of network programming techniques, from server-side and client-side solutions to advanced client-server communication models; reviews network-based data storage and multimedia transfer; includes an extensive set of practical code examples, together with detailed comments and explanations. This comprehensive and authoritative guide is an invaluable asset for all researchers interested in computer networking, whether they wish to understand the underlying architectures and paradigms, or to obtain useful advice on building communication-based programs. Advanced undergraduate and postgraduate students will also find the book to be an excellent supplementary textbook for modules on network programming.
This volume contains the papers presented. at the Third IFIP International Working Conference on Dependable Computing for Critical Applications, sponsored by IFIP Working Group 10.4 and held in Mondello (Sicily), Italy on September 14-16, 1992. System developers increasingly apply computers where they can affect the safety and security of people and equipment. The Third IFIP International Working Conference on Dependable Computing for Critical Applications, like its predecessors, addressed various aspects of computer system dependability, a broad term defined as the degree of trust that may justifiably be placed in a system's reliability, availability, safety, security, and performance. Because the scope of the conference was so broad, we hope the presentations and discussions will contribute to the integration of these concepts so that future computer-based systems will indeed be more dependable. The Program Committee selected 18 papers for presentation from a total of 7 4 submissions at a May meeting in Newcastle upon Tyne, UK. The resulting program represented a broad spectrum of interests, with papers from universities, corporations, and government agencies in eight countries. Much diligent work by the Program Committee and the quality of reviews from more than a hundred external referees from around the world, for which we are most grateful, significantly eased the production of this technical program.
This book constitutes the thoroughly refereed post-conference proceedings of the Second International Joint Conference on Knowledge Discovery, Knowledge Engineering, and Knowledge Management, IC3K 2010, held in Valencia, Spain, in October 2010. This book includes revised and extended versions of a strict selection of the best papers presented at the conference; 26 revised full papers together with 2 invited lectures were carefully reviewed and selected from 369 submissions. According to the three covered conferences KDIR 2010, KEOD 2010, and KMIS 2010, the papers are organized in topical sections on knowledge discovery and information retrieval, knowledge engineering and ontology development, and on knowledge management and information sharing.
At the beginning of the 1990s research started in how to combine soft comput ing with reconfigurable hardware in a quite unique way. One of the methods that was developed has been called evolvable hardware. Thanks to evolution ary algorithms researchers have started to evolve electronic circuits routinely. A number of interesting circuits - with features unreachable by means of con ventional techniques - have been developed. Evolvable hardware is quite pop ular right now; more than fifty research groups are spread out over the world. Evolvable hardware has become a part of the curriculum at some universi ties. Evolvable hardware is being commercialized and there are specialized conferences devoted to evolvable hardware. On the other hand, surprisingly, we can feel the lack of a theoretical background and consistent design methodology in the area. Furthermore, it is quite difficult to implement really innovative and practically successful evolvable systems using contemporary digital reconfigurable technology."
Business Component-Based Software Engineering, an edited volume, aims to complement some other reputable books on CBSE, by stressing how components are built for large-scale applications, within dedicated development processes and for easy and direct combination. This book will emphasize these three facets and will offer a complete overview of some recent progresses. Projects and works explained herein will prompt graduate students, academics, software engineers, project managers and developers to adopt and to apply new component development methods gained from and validated by the authors. The authors of Business Component-Based Software Engineering are academic and professionals, experts in the field, who will introduce the state of the art on CBSE from their shared experience by working on the same projects. Business Component-Based Software Engineering is designed to meet the needs of practitioners and researchers in industry, and graduate-level students in Computer Science and Engineering.
Real-time computing systems are vital to a wide range of applications. For example, they are used in the control of nuclear reactors and automated manufacturing facilities, in controlling and tracking air traffic, and in communication systems. In recent years, real-time systems have also grown larger and become more critical. For instance, advanced aircraft such as the space shuttle must depend heavily on computer sys tems Carlow 84]. The centralized control of manufacturing facilities and assembly plants operated by robots are other examples at the heart of which lie embedded real-time systems. Military defense systems deployed in the air, on the ocean surface, land and underwater, have also been increasingly relying upon real-time systems for monitoring and operational safety purposes, and for retaliatory and containment measures. In telecommunications and in multi-media applications, real time characteristics are essential to maintain the integrity of transmitted data, audio and video signals. Many of these systems control, monitor or perform critical operations, and must respond quickly to emergency events in a wide range of embedded applications. They are therefore required to process tasks with stringent timing requirements and must perform these tasks in a way that these timing requirements are guaranteed to be met. Real-time scheduling al gorithms attempt to ensure that system timing behavior meets its specifications, but typically assume that tasks do not share logical or physical resources. Since resource-sharing cannot be eliminated, synchronization primitives must be used to ensure that resource consis tency constraints are not violated."
This book constitutes the thoroughly refereed post-conference proceedings of the 9th International Symposium on Computer Music Modeling and Retrieval, CMMR 2012, held in London, UK, in June 2012. The 28 revised full papers presented were carefully reviewed and selected for inclusion in this volume. The papers have been organized in the following topical sections: music emotion analysis; 3D audio and sound synthesis; computer models of music perception and cognition; music emotion recognition; music information retrieval; film soundtrack and music recommendation; and computational musicology and music education. The volume also includes selected papers from the Cross-Disciplinary Perspectives on Expressive Performance Workshop held within the framework of CMMR 2012.
Research into grid computing has been driven by the need to solve large-scale, increasingly complex problems for scientific applications. Yet the applications of grid computing for business and casual users did not begin to emerge until the development of the concept of cloud computing, fueled by advances in virtualization techniques, coupled with the increased availability of ever-greater Internet bandwidth. The appeal of this new paradigm is mainly based on its simplicity, and the affordable price for seamless access to both computational and storage resources. This timely text/reference introduces the fundamental principles and techniques underlying grids, clouds and virtualization technologies, as well as reviewing the latest research and expected future developments in the field. Readers are guided through the key topics by internationally recognized experts, enabling them to develop their understanding of an area likely to play an ever more significant role in coming years. Topics and features: presents contributions from an international selection of experts in the field; provides a thorough introduction and overview of existing technologies in grids, clouds and virtualization, including a brief history of the field; examines the basic requirements for performance isolation of virtual machines on multi-core servers, analyzing a selection of system virtualization technologies; examines both business and scientific applications of grids and clouds, including their use in the life sciences and for high-performance computing; explores cloud building technologies, architectures for enhancing grid infrastructures with cloud computing, and cloud performance; discusses energy aware grids and clouds, workflows on grids and clouds, and cloud and grid programming models. This useful text will enable interested readers to familiarize themselves with the key topics of grids, clouds and virtualization, and to contribute to new advances in the field. Researchers, undergraduate and graduate students, system designers and programmers, and IT policy makers will all benefit from the material covered.
The process of designing large real-time embedded signal processing systems is plagued by a lack of coherent specification and design methodology. A canonical waterfall design process is commonly used to specify, design, and implement these systems with commercial-off-the-shelf (COTS) multiprocessing (MP) hardware and software. Powerful frameworks exist for each individual phase of this canonical design process, but no single methodology exists which enables these frameworks to work together coherently, i.e., allowing the output of a framework used in one phase to be consumed by a different framework used in the next phase. This lack of coherence usually leads to design errors that are not caught until well in to the implementation phase. Since the cost of redesign increases as the design moves through these three stages, redesign is the most expensive if not performed until the implementation phase, thus making the current incoherent methodology costly.Specification and Design Methodology for Real-Time Embedded Systems shows how designs targeting COTS MP technologies can be improved by providing a coherent coupling between these frameworks, a quality known as model continuity. This book presents a new specification and design methodology (SDM) which accomplishes the requirements specification, design exploration, and implementation of COTS MP-based signal processing systems by using powerful commercial frameworks that are intelligently integrated into a single domain-specific SDM. From the foreword: This book is remarkably practical. It provides an excellent snapshot of the state-of-the-art and gives the reader a good understanding of both the fundamental challenges of specification and design as well as a unified and quantified ability to assess a given methodology. Daniel Gajski, University of California
This book constitutes the refereed proceedings of the 9th International Workshop on Economics of Grids, Clouds, Systems, and Services, GECON 2012, held in Berlin, Germany, in November 2012. The 12 revised full papers presented together with 6 work in progress papers were carefully reviewed and selected from more than 36 submissions. The papers are organized in the following topical sections: market mechanisms, pricing and negotiation; resource allocation, scheduling and admission control; work in progress on tools and techniques for cost-efficient service selection; market modeling; trust; cloud computing in education; and work in progress on cloud adoption and business models.
This two volumes set LNAI 8102 and LNAI 8103 constitutes the refereed proceedings of the 6th International Conference on Intelligent Robotics and Applications, ICIRA 2013, held in Busan, South Korea, in September 2013. The 147 revised full papers presented were carefully reviewed and selected from 184 submissions. The papers discuss various topics from intelligent robotics, automation and mechatronics with particular emphasis on technical challenges associated with varied applications such as biomedical application, industrial automation, surveillance and sustainable mobility.
Challenges in Design and Implementation of Middlewares for Real-Time Systems brings together in one place important contributions and up-to-date research results in this fast moving area. Challenges in Design and Implementation of Middlewares for Real-Time Systems serves as an excellent reference, providing insight into some of the most challenging research issues in the field.
This book constitutes the refereed proceedings of the International Conference on Data and Knowledge Engineering, ICDKE 2012, held in Wuyishan, Fujian, China, in November 2012. The conference was co-located with the 6th International Conference on Network and System Security, NSS 2012. The 13 revised full papers of ICDKE 2012 were carefully reviewed and selected from 53 submissions. The papers cover the following topics: artificial intelligence and data engineering; knowledge discovery and data management; information extraction and retrieval and data security.
Real-time systems are defined as those for which correctness depends not only on the logical properties of the produced results, but also on the temporal properties of these results. In a database, real-time means that in addition to typical logical consistency constraints, such as a constraint on a data item's value, there are constraints on when transactions execute and on the `freshness' of the data transactions access. The challenges and tradeoffs faced by the designers of real-time database systems are quite different from those faced by the designers of general-purpose database systems. To achieve the fundamental requirements of timeliness and predictability, not only do conventional methods for scheduling and transaction management have to be redesigned, but also new concepts that have not been considered in conventional database systems or in real-time systems need to be added. Real-Time Database and Information Systems: Research Advances is devoted to new techniques for scheduling of transactions, concurrency management, transaction logging, database languages, and new distributed database architectures. Real-Time Database and Information Systems: Research Advances is primarily intended for practicing engineers and researchers working in the growing area of real-time database and information retrieval systems. For practitioners, the book will provide a much needed bridge for technology transfer and continued education. For researchers, the book will provide a comprehensive reference for well-established results. The book can also be used in a senior or graduate level course on real-time systems, real-time database systems, and database systems, or closely related courses.
Vorwort In der Natur entwickelten sich die Echtzeitsysteme seit einigen 100 Mil- Honen Jahren. Tierische Nervensysteme haben zur Aufgabe, auf die Nachrichten aus der Umwelt die Steuerungsbefehle an die aktiven Or- gane zu geben. Dabei spielen zum Beispiel bedingte Reflexe eine wichtige Rolle. Vielleicht kann man die Entstehung des Menschen etwa zu der Zeit ansetzen, als sein sich allmahlich entwickelndes Gehirn Gedanken entwickelte, deren Bedeutung in vorausplanender Weise iiber die gerade vorliegende Situation hinausging. Das fiihrte schliesslich unter anderem zum heutigen Wissenschaftler, der seine Theorien und Systeme aufgrund langwieriger Uberlegungen aufbaut. Die Entwicklung der Computer ging im wesentlichen den umgekehrten Weg. Zunachst diente sie nur der Durchfiihrung "starrer" Programme, wie z.B. das erste programmgesteuerte Rechengerat Z3, das der Unterzeichner im Jahre 1941 vorfiihren konnte. Es folgte unter an- derem ein Spezialgerat zur Fliigelvermessung, das man als den ersten Prozessrechner bezeichnen kann. Es wurden etwa vierzig als Analog- Digital-Wandler arbeitende Messuhren yom Rechnerautomaten abgele- sen und im Rahmen eines Programms als Variable verarbeitet. Abel' auch das erfolgte noch in starrer Reihenfolge. Die echte Prozesssteuerung - heute auch Echtzeitsysteme genannt - erfordert aber ein Reagieren auf bestandig wechselnde Situationen.
This book constitutes the refereed proceedings of the 24th IFIP WG 6.1 International Conference on Testing Software and Systems, ICTSS 2012, held in Aalborg, Denmark, in November 2012. The 16 revised full papers presented together with 2 invited talks were carefully selected from 48 submissions. The papers are organized in topical sections on testing in practice, test frameworks for distributed systems, testing of embedded systems, test optimization, and new testing methods.
Despite the growing interest in Real-Time Database Systems, there is no single book that acts as a reference to academics, professionals, and practitioners who wish to understand the issues involved in the design and development of RTDBS. Real-Time Database Systems: Issues and Applications fulfills this need. This book presents the spectrum of issues that may arise in various real-time database applications, the available solutions and technologies that may be used to address these issues, and the open problems that need to be tackled in the future. With rapid advances in this area, several concepts have been proposed without a widely accepted consensus on their definitions and implications. To address this need, the first chapter is an introduction to the key RTDBS concepts and definitions, which is followed by a survey of the state of the art in RTDBS research and practice. The remainder of the book consists of four sections: models and paradigms, applications and benchmarks, scheduling and concurrency control, and experimental systems. The chapters in each section are contributed by experts in the respective areas. Real-Time Database Systems: Issues and Applications is primarily intended for practicing engineers and researchers working in the growing area of real-time database systems. For practitioners, the book will provide a much needed bridge for technology transfer and continued education. For researchers, this book will provide a comprehensive reference for well-established results. This book can also be used in a senior or graduate level course on real-time systems, real-time database systems, and database systems or closely related courses.
This book is a compilation of research accomplishments in the fields of modeling, simulation, and their applications, as presented at AsiaSim 2011 (Asia Simulation Conference 2011). The conference, held in Seoul, Korea, November 16-18, was organized by ASIASIM (Federation of Asian Simulation Societies), KSS (Korea Society for Simulation), CASS (Chinese Association for System Simulation), and JSST (Japan Society for Simulation Technology). AsiaSim 2011 provided a forum for scientists, academicians, and professionals from the Asia-Pacific region and other parts of the world to share their latest exciting research findings in modeling and simulation methodologies, techniques, and their tools and applications in military, communication network, industry, and general engineering problems.
Formal Techniques in Real-Time and Fault-Tolerant Systems focuses on the state of the art in formal specification, development and verification of fault-tolerant computing systems. The term `fault-tolerance' refers to a system having properties which enable it to deliver its specified function despite (certain) faults of its subsystem. Fault-tolerance is achieved by adding extra hardware and/or software which corrects the effects of faults. In this sense, a system can be called fault-tolerant if it can be proved that the resulting (extended) system under some model of reliability meets the reliability requirements. The main theme of Formal Techniques in Real-Time and Fault-Tolerant Systems can be formulated as follows: how do the specification, development and verification of conventional and fault-tolerant systems differ? How do the notations, methodology and tools used in design and development of fault-tolerant and conventional systems differ? Formal Techniques in Real-Time and Fault-Tolerant Systems is divided into two parts. The chapters in Part One set the stage for what follows by defining the basic notions and practices of the field of design and specification of fault-tolerant systems. The chapters in Part Two represent the `how-to' section, containing examples of the use of formal methods in specification and development of fault-tolerant systems. The book serves as an excellent reference for researchers in both academia and industry, and may be used as a text for advanced courses on the subject.
With the rapid growth of networking and high-computing power, the demand for large-scale and complex software systems has increased dramatically. Many of the software systems support or supplant human control of safety-critical systems such as flight control systems, space shuttle control systems, aircraft avionics control systems, robotics, patient monitoring systems, nuclear power plant control systems, and so on. Failure of safety-critical systems could result in great disasters and loss of human life. Therefore, software used for safety critical systems should preserve high assurance properties. In order to comply with high assurance properties, a safety-critical system often shares resources between multiple concurrently active computing agents and must meet rigid real-time constraints. However, concurrency and timing constraints make the development of a safety-critical system much more error prone and arduous. The correctness of software systems nowadays depends mainly on the work of testing and debugging. Testing and debugging involve the process of de tecting, locating, analyzing, isolating, and correcting suspected faults using the runtime information of a system. However, testing and debugging are not sufficient to prove the correctness of a safety-critical system. In contrast, static analysis is supported by formalisms to specify the system precisely. Formal verification methods are then applied to prove the logical correctness of the system with respect to the specification. Formal verifica tion gives us greater confidence that safety-critical systems meet the desired assurance properties in order to avoid disastrous consequences.
Sir Isaac Newton's philosophi Naturalis Principia Mathematica'(the Principia) contains a prose-style mixture of geometric and limit reasoning that has often been viewed as logically vague. In A Combination of Geometry Theorem Proving and Nonstandard Analysis, Jacques Fleuriot presents a formalization of Lemmas and Propositions from the Principia using a combination of methods from geometry and nonstandard analysis. The mechanization of the procedures, which respects much of Newton's original reasoning, is developed within the theorem prover Isabelle. The application of this framework to the mechanization of elementary real analysis using nonstandard techniques is also discussed.
The International Workshop on "The Internet Challenge: Technology and Applications" is the fifth in a successful series of workshops that were established by Shanghai Jiao Tong University and Technische Universitat Berlin. The goal of those workshops is to bring together researchers from both universities in order to exchange research results achieved in common projects of the two partner universities or to present interesting new work that might lead to new cooperation. The series of workshops started in 1990 with the "International Workshop on Artificial Intelligence" and was continued with the "International Workshop on Advanced Software Technology" in 1994. Both workshops have been hosted by Shanghai Jiao Tong University. In 1998 the third workshop took place in Berlin. This "International Workshop on Communication Based Systems" was essentially based on results from the Graduiertenkolleg on Communication Based systems that was funded by the German Research Society (DFG) from 1991 to 2000. The fourth "International Workshop on Robotics and its Applications" was held in Shanghai in 2000 supported by VDIIVDE-GMA and GI.
This volume contains a selection of papers that focus on the state-of the-art in formal specification and verification of real-time computing systems. Preliminary versions of these papers were presented at a workshop on the foundations of real-time computing sponsored by the Office of Naval Research in October, 1990 in Washington, D. C. A companion volume by the title Foundations of Real-Time Computing: Scheduling and Resource Management complements this hook by addressing many of the recently devised techniques and approaches for scheduling tasks and managing resources in real-time systems. Together, these two texts provide a comprehensive snapshot of current insights into the process of designing and building real time computing systems on a scientific basis. The notion of real-time system has alternative interpretations, not all of which are intended usages in this collection of papers. Different communities of researchers variously use the term real-time to refer to either very fast computing, or immediate on-line data acquisition, or deadline-driven computing. This text is concerned with the formal specification and verification of computer software and systems whose correct performance is dependent on carefully orchestrated interactions with time, e. g., meeting deadlines and synchronizing with clocks. Such systems have been enabled for a rapidly increasing set of diverse end-uses by the unremitting advances in computing power per constant-dollar cost and per constant-unit-volume of space. End use applications of real-time computers span a spectrum that includes transportation systems, robotics and manufacturing, aerospace and defense, industrial process control, and telecommunications."
This book constitutes the refereed proceedings of the International Conference on Multiscore Software Engineering, Performance, and Tools, MUSEPAT 2013, held in Saint Petersburg, Russia, in August 2013. The 9 revised papers were carefully reviewed and selected from 25 submissions. The accepted papers are organized into three main sessions and cover topics such as software engineering for multicore systems; specification, modeling and design; programing models, languages, compiler techniques and development tools; verification, testing, analysis, debugging and performance tuning, security testing; software maintenance and evolution; multicore software issues in scientific computing, embedded and mobile systems; energy-efficient computing as well as experience reports. |
![]() ![]() You may like...
Research Anthology on Artificial Neural…
Information R Management Association
Hardcover
R14,040
Discovery Miles 140 400
Research Anthology on Artificial Neural…
Information R Management Association
Hardcover
R14,034
Discovery Miles 140 340
Recent Trends in Computational…
Siddhartha Bhattacharyya, Paramartha Dutta, …
Paperback
R3,700
Discovery Miles 37 000
Handbook of Research on Advanced…
Siddhartha Bhattacharyya, Pinaki Banerjee, …
Hardcover
R7,638
Discovery Miles 76 380
Research Anthology on Artificial Neural…
Information R Management Association
Hardcover
R14,050
Discovery Miles 140 500
Deep Learning Applications for…
Monica R. Mundada, Seema S., …
Hardcover
R7,211
Discovery Miles 72 110
|