![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Artificial intelligence > Knowledge-based systems / expert systems
This book is the first in aseries on novellow power design architectures, methods and design practices. It results from of a large European project started in 1997, whose goal is to promote the further development and the faster and wider industrial use of advanced design methods for reducing the power consumption of electronic systems. Low power design became crucial with the wide spread of portable information and cornrnunication terminals, where a small battery has to last for a long period. High performance electronics, in addition, suffers from a permanent increase of the dissipated power per square millimetre of silicon, due to the increasing eIock-rates, which causes cooling and reliability problems or otherwise limits the performance. The European Union's Information Technologies Programme 'Esprit' did there fore launch a 'Pilot action for Low Power Design', wh ich eventually grew to 19 R&D projects and one coordination project, with an overall budget of 14 million Euro. It is meanwhile known as European Low Power Initiative for Electronic System Design (ESD-LPD) and will be completed by the end of 2001. It involves 30 major Euro pean companies and 20 well-known institutes. The R&D projects aims to develop or demonstrate new design methods for power reduction, while the coordination project takes care that the methods, experiences and results are properly documented and pub licised."
This book is the latest contribution to the Chip Design Languages series and it consists of selected papers presented at the Forum on Specifications and Design Languages (FDL'06), in September 2006. The book represents the state-of-the-art in research and practice, and it identifies new research directions. It highlights the role of specification and modelling languages, and presents practical experiences with specification and modelling languages.
This book proposes novel memory hierarchies and software optimization techniques for the optimal utilization of memory hierarchies. It presents a wide range of optimizations, progressively increasing in the complexity of analysis and of memory hierarchies. The final chapter covers optimization techniques for applications consisting of multiple processes found in most modern embedded devices.
This book constitutes the refereed proceedings of Industry Oriented Conferences held at IFIP 20th World Computer Congress in September 2008. The IFIP series publishes state-of-the-art results in the sciences and technologies of information and communication. The scope of the series includes: foundations of computer science; software theory and practice; education; computer applications in technology; communication systems; systems modeling and optimization; information systems; computers and society; computer systems technology; security and protection in information processing systems; artificial intelligence; and human-computer interaction. Proceedings and post-proceedings of refereed international conferences in computer science and interdisciplinary fields are featured. These results often precede journal publication and represent the most current research. The principal aim of the IFIP series is to encourage education and the dissemination and exchange of information about all aspects of computing.
This book offers up a deep understanding of concepts and practices behind the composition of heterogeneous components. After the analysis of existing computation and execution models used for the specification and validation of different sub-systems, the book introduces a systematic approach to build an execution model for systems composed of heterogeneous components. Mixed continuous/discrete and hardware/software systems are used to illustrate these concepts. The benefit of reading this book is to arrive at a clear vision of the theory and practice of specification and validation of complex modern systems. Numerous examples give designers highly applicable solutions.
Increasing complexity of modern embedded systems demands system designers to ramp up their design productivity without compromising performance goals. This is promoted by modern Electronic System Level (ESL) techniques. Language-driven Exploration and Implementation of Partially Re-configurable ASIPs addresses an important segment of the ESL area by modeling partially re-configurable processors via high-level Architecture Description Language (ADL). This approach also hints an imminent evolution in the area of re-configurable system design.
This book presents a framework for mobile information systems, focusing on quality of service and adaptability at all architectural levels. These levels range from adaptive applications to e-services, middleware, and infrastructural elements, as developed in the "Multichannel Adaptive Information Systems" (MAIS) project. The design models, methods, and tools developed in the project allow the realization of adaptive mobile information systems in a variety of different architectures.
Hopping, climbing and swimming robots, nano-size neural networks, motorless walkers, slime mould and chemical brains - "Artificial Life Models in Hardware" offers unique designs and prototypes of life-like creatures in conventional hardware and hybrid bio-silicon systems. Ideas and implementations of living phenomena in non-living substrates cast a colourful picture of state-of-art advances in hardware models of artificial life.
This book brings Network Calculus closer to the network professional and will also have real appeal for postgraduates studying network performance. It provides valuable analytical tools and uses J as a means of providing a practical treatment of the subject. It builds a bridge between mathematics theory and the practical use of computers in the field of network performance analysis.
Most innovations in the car industry are based on software and electronics, and IT will soon constitute the major production cost factor. It seems almost certain that embedded IT security will be crucial for the next generation of applications. Yet whereas software safety has become a relatively well-established field, the protection of automotive IT systems against manipulation or intrusion has only recently started to emerge. Lemke, Paar, and Wolf collect in this volume a state-of-the-art overview on all aspects relevant for IT security in automotive applications. After an introductory chapter written by the editors themselves, the contributions from experienced experts of different disciplines are structured into three parts. "Security in the Automotive Domain" describes applications for which IT security is crucial, like immobilizers, tachographs, and software updates. "Embedded Security Technologies" details security technologies relevant for automotive applications, e.g., symmetric and asymmetric cryptography, and wireless security. "Business Aspects of IT Systems in Cars" shows the need for embedded security in novel applications like location-based navigation systems and personalization. The first book in this area of fast-growing economic and scientific importance, it is indispensable for both researchers in software or embedded security and professionals in the automotive industry.
Pervasive healthcare is the conceptual system of providing healthcare to anyone, at anytime, and anywhere by removing restraints of time and location while increasing both the coverage and the quality of healthcare. Pervasive Healthcare Computing is at the forefront of this research, and presents the ways in which mobile and wireless technologies can be used to implement the vision of pervasive healthcare. This vision includes prevention, healthcare maintenance and checkups; short-term monitoring (home healthcare), long-term monitoring (nursing home), and personalized healthcare monitoring; and incidence detection and management, emergency intervention, transportation and treatment. The pervasive healthcare applications include intelligent emergency management system, pervasive healthcare data access, and ubiquitous mobile telemedicine. Pervasive Healthcare Computing includes the treatment of several new wireless technologies and the ways in which they will implement the vision of pervasive healthcare.
Here is an extremely useful book that provides insight into a number of different flavors of processor architectures and their design, software tool generation, implementation, and verification. After a brief introduction to processor architectures and how processor designers have sometimes failed to deliver what was expected, the authors introduce a generic flow for embedded on-chip processor design and start to explore the vast design space of on-chip processing. The authors cover a number of different types of processor core.
This essential resource for professionals and advanced students in security programming and system design introduces the foundations of programming systems security and the theory behind access control models, and addresses emerging access control mechanisms.
Thiseditedbookispublishedin honorofDr. GeorgeJ. Vachtsevanos, ourDr. V, c- rently Professor Emeritus, School of Electrical and Computer Engineering, Georgia Institute of Technology, on the occasion of his 70th birthday and for his more than 30 years of contribution to the discipline of Intelligent Control and its application to a wide spectrum of engineering and bioengineering systems. The book is nothing but a very small token of appreciation from Dr. V's former graduate students, his peers and colleagues in the profession - and not only - to the Scientist, the Engineer, the Professor, the mentor, but most important of all, to the friend and human being. All those who have met Dr. V over the years and haveinteractedwith himin someprofessionaland/orsocial capacityunderstandthis statement: Georgenevermadeanybodyfeelinferiortohim, hehelpedandsupported everybody, and he was there when anybody needed him I was not Dr. V's student. I rst met him and his wife Athena more than 26 years ago during one of their visits to RPI, in the house of my late advisor, Dr. George N. Saridis. Since then, I have been very fortunate to have had and continue to have interactions with him. It is not an exaggeration if I say that we all learned a lot from him.
With the rapid advances in technology, the conventional academic and research departments of Electronics engineering, Electrical Engineering, Computer Science, Instrumentation Engineering over the globe are forced to come together and update their curriculum with few common interdisciplinary courses in order to come out with the engineers and researchers with muli-dimensional capabilities. The gr- ing perception of the 'Hardware becoming Soft' and 'Software becoming Hard' with the emergence of the FPGAs has made its impact on both the hardware and software professionals to change their mindset of working in narrow domains. An interdisciplinary field where 'Hardware meets the Software' for undertaking se- ingly unfeasible tasks is System on Chip (SoC) which has become the basic pl- form of modern electronic appliances. If it wasn't for SoCs, we wouldn't be driving our car with foresight of the traffic congestion before hand using GPS. Without the omnipresence of the SoCs in our every walks of life, the society is wouldn't have evidenced the rich benefits of the convergence of the technologies such as audio, video, mobile, IPTV just to name a few. The growing expectations of the consumers have placed the field of SoC design at the heart of at variance trends. On one hand there are challenges owing to design complexities with the emergence of the new processors, RTOS, software protocol stacks, buses, while the brutal forces of deep submicron effects such as crosstalk, electromigration, timing closures are challe- ing the design metrics.
This book analyzes the causes of failures in computing systems, their consequences, as weIl as the existing solutions to manage them. The domain is tackled in a progressive and educational manner with two objectives: 1. The mastering of the basics of dependability domain at system level, that is to say independently ofthe technology used (hardware or software) and of the domain of application. 2. The understanding of the fundamental techniques available to prevent, to remove, to tolerate, and to forecast faults in hardware and software technologies. The first objective leads to the presentation of the general problem, the fault models and degradation mechanisms wh ich are at the origin of the failures, and finally the methods and techniques which permit the faults to be prevented, removed or tolerated. This study concerns logical systems in general, independently of the hardware and software technologies put in place. This knowledge is indispensable for two reasons: * A large part of a product' s development is independent of the technological means (expression of requirements, specification and most of the design stage). Very often, the development team does not possess this basic knowledge; hence, the dependability requirements are considered uniquely during the technological implementation. Such an approach is expensive and inefficient. Indeed, the removal of a preliminary design fault can be very difficult (if possible) if this fault is detected during the product's final testing.
In recent years, increases in the amount and changes in the distribution of air traffic have been very dramatic and are continuing. The need for changes in the current air traffic systems is equally clear. While automation is generally accepted as a method of improving system safety and performance, high levels of automation in complex human-machine systems can have a negative effect on total system performance and have been identified as contributing factors in many accidents and failures. Those responsible for designing the advanced air traffic control systems to be implemented throughout the alliance during the next decade need to be aware of recent progress concerning the most effective application of automation and artificial intelligence in human-computer systems. This volume gives the proceedings of the NATO Advanced Study Institute held in Maratea, Italy, June 18-29, 1990, at which these issues were discussed.
W.J.Quirk 1.1 Real-time software and the real world Real-time software and the real world are inseparably related. Real time cannot be turned back and the real world will not always forget its history. The consequences of previous influences may last for a long time and the undesired effects may range from being inconvenient to disastrous in both economic and human terms. As a result, there is much pressure to develop and apply techniques to improve the reliability of real-time software so that the frequency and consequences of failure are reduced to a level that is as low as reasonably achievable. This report is about such techniques. After a detailed description of the software life cycle, a chapter is devoted to each of the four principle categories of technique available at present. These cover all stages of the software development process and each chapter identifies relevant techniques, the stages to which they are applicable and their effectiveness in improving real-time software reliability. 1.2 The characteristics of real-time software As well as the enhanced reliability requirement discussed above, real-time software has a number of other distinguishing characteristics. First, the sequencing and timing of inputs are determined by the real world and not by the programmer. Thus the program needs to be prepared for the unexpected and the demands made on the system may be conflicting. Second, the demands on the system may occur in parallel rather than in sequence.
This book constitutes the proceedings of the 13th International Symposium on Stabilization, Safety, and Security of Distributed Systems, SSS 2011, held in Grenoble, France, in October 2011. The 29 papers presented were carefully reviewed and selected from 79 submissions. They cover the following areas: ad-hoc, sensor, and peer-to-peer networks; safety and verification; security; self-organizing and autonomic systems; and self-stabilization.
The two-volume set LNCS 6773-6774 constitutes the refereed proceedings of the International Conference on Virtual and Mixed Reality 2011, held as Part of HCI International 2011, in Orlando, FL, USA, in July 2011, jointly with 10 other conferences addressing the latest research and development efforts and highlighting the human aspects of design and use of computing systems. The 43 revised papers included in the first volume were carefully reviewed and selected from numerous submissions. The papers are organized in the following topical sections: augmented reality applications; virtual and immersive environments; novel interaction devices and techniques in VR; human physiology and behavior in VR environments.
This book constitutes thoroughly refereed post-conference proceedings of the workshops of the 16th International Conference on Parallel Computing, Euro-Par 2010, held in Ischia, Italy, in August/September 2010. The papers of these 9 workshops HeteroPar, HPCC, HiBB, CoreGrid, UCHPC, HPCF, PROPER, CCPI, and VHPC focus on promotion and advancement of all aspects of parallel and distributed computing.
This two-volume set LNCS 6771 and 6772 constitutes the refereed proceedings of the Symposium on Human Interface 2011, held in Orlando, FL, USA in July 2011 in the framework of the 14th International Conference on Human-Computer Interaction, HCII 2011 with 10 other thematically similar conferences. The 137 revised papers presented in the two volumes were carefully reviewed and selected from numerous submissions. The papers accepted for presentation thoroughly cover the thematic area of human interface and the management of information. The 75 papers of this first volume address the following major topics: design and development methods and tools; information and user interfaces design; visualisation techniques and applications; security and privacy; touch and gesture interfaces; adaption and personalisation; and measuring and recognising human behavior.
This book constitutes the proceedings of the 19th International Conference on Conceptual Structures, ICCS 2011, held in Derby, UK, in July 2011. The 18 full papers and 4 short papers presented together with 12 workshop papers were carefully reviewed and selected for inclusion in the book. The volume also contains 3 invited talks. ICCS focuses on the useful representation and analysis of conceptual knowledge with research and business applications. It advances the theory and practice in connecting the user's conceptual approach to problem solving with the formal structures that computer applications need to bring their productivity to bear. Conceptual structures (CS) represent a family of approaches that builds on the successes of artificial intelligence, business intelligence, computational linguistics, conceptual modelling, information and Web technologies, user modelling, and knowledge management. Two of the workshops contained in this volume cover CS and knowledge discovery in under-traversed domains and in task specific information retrieval. The third addresses CD in learning, teaching and assessment.
Innovation in Manufacturing Networks A fundamental concept of the emergent business, scientific and technological paradigms ces area, innovation the ability to apply new ideas to products, processes, organizational practices and business models - is crucial for the future competitiveness of organizations in a continually increasingly globalised, knowledge-intensive marketplace. Responsiveness, agility as well as the high performance of manufacturing systems is responsible for the recent changes in addition to the call for new approaches to achieve cost-effective responsiveness at all the levels of an enterprise. Moreover, creating appropriate frameworks for exploring the most effective synergies between human potential and automated systems represents an enormous challenge in terms of processes characterization, modelling, and the development of adequate support tools. The implementation and use of Automation Systems requires an ever increasing knowledge of enabling technologies and Business Practices. Moreover, the digital and networked world will surely trigger new business practices. In this context and in order to achieve the desired effective and efficiency performance levels, it is crucial to maintain a balance between both the technical aspects and the human and social aspects when developing and applying new innovations and innovative enabling technologies. BASYS conferences have been developed and organized so as to promote the development of balanced automation systems in an attempt to address the majority of the current open issues.
The Second International Conference on High-Performance Computing and Appli- tions (HPCA 2009) was a follow-up event of the successful HPCA 2004. It was held in Shanghai, a beautiful, active, and modern city in China, August 10-12, 2009. It served as a forum to present current work by researchers and software developers from around the world as well as to highlight activities in the high-performance c- puting area. It aimed to bring together research scientists, application pioneers, and software developers to discuss problems and solutions and to identify new issues in this area. This conference emphasized the development and study of novel approaches for high-performance computing, the design and analysis of high-performance - merical algorithms, and their scientific, engineering, and industrial applications. It offered the conference participants a great opportunity to exchange the latest research results, heighten international collaboration, and discuss future research ideas in HPCA. In addition to 24 invited presentations, the conference received over 300 contr- uted submissions from over ten countries and regions worldwide, about 70 of which were accepted for presentation at HPCA 2009. The conference proceedings contain some of the invited presentations and contributed submissions, and cover such research areas of interest as numerical algorithms and solutions, high-performance and grid c- puting, novel approaches to high-performance computing, massive data storage and processing, hardware acceleration, and their wide applications. |
You may like...
|