![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer hardware & operating systems > Operating systems & graphical user interfaces (GUIs) > General
Today, more and more Web sites are providing content in multiple
languages for targeted countries, and more and more products are
being designed for cultural differences in mind. However, the
concept of cross-cultural design has not yet become a strong force
in the practitioners' and educators' agenda. This book looks at
techniques, software, tools, the current state of the art, and
future directions that one needs to understand for a successful
application of cross-cultural interface design.
This book constitutes the thoroughly refereed post-conference proceedings of the Second International Workshop on Smart Grid Security, SmartGridSec 2014, held in Munich, Germany, in February 2014. The volume contains twelve corrected and extended papers presented at the workshop which have undergone two rounds of reviewing and improvement. The engineering, deployment and operation of the future Smart Grid will be an enormous project that will require the active participation of many stakeholders with different interests and views regarding the security and privacy goals, technologies, and solutions. There is an increasing need for workshops that bring together researchers from different communities, from academia and industry, to discuss open research topics in the area of future Smart Grid security.
This book constitutes the refereed proceedings of the 11th International Conference on Wired/Wireless Internet Communications, WWIC 2013, held in St. Petersburg, Russia, during June 5-7, 2013. The 21 revised full papers presented were carefully reviewed and selected from numerous submissions. The papers are organized in topical sections on: MAC and scheduling; mobility; fixed networks; services and security; sensor networks; services; and wireless.
The engineering, deployment and security of the future smart grid will be an enormous project requiring the consensus of many stakeholders with different views on the security and privacy requirements, not to mention methods and solutions. The fragmentation of research agendas and proposed approaches or solutions for securing the future smart grid becomes apparent observing the results from different projects, standards, committees, etc, in different countries. The different approaches and views of the papers in this collection also witness this fragmentation. This book contains three full-paper length invited papers and 7 corrected and extended papers from the First International Workshop on Smart Grid Security, SmartGridSec 2012, which brought together researchers from different communities from academia and industry in the area of securing the Future Smart Grid and was held in Berlin, Germany, on December 3, 2012.
This book constitutes the thoroughly refereed post-conference proceedings of the 17th International Workshop on Job Scheduling Strategies for Parallel Processing, JSSPP 2013, held Boston, MA, USA, in May 2013. The 10 revised papers presented were carefully reviewed and selected from 20 submissions. The papers cover the following topics parallel scheduling for commercial environments, scientific computing, supercomputing and cluster platforms.
In modern scientific investigation the fields of biochemistry, molecular biology, and morphology comprise an indivisible area of study. The present book results from the cooperation of a bioehernist and morphologists: the revision and unified treatment of available data is the primary object of our work. A comprehensive review of all the available literature is therefore beyond the scope of this volume. lt is intended to be a manual to be used in the laboratory, with convenient guidelines for practical work. Plant microbodies have been treated by B. GERHARDT in Volume 5 of this series. The discovery of fatty acid ss-oxidation in animal peroxisomes has proved once more that plant and animal microbodies are members of the same family of organelles. lt provided new insights into the physiological meaning of these particles; our understanding of these "classical" cell organelies is undergoing continual alteration and development. PETER BoecK Vienna, July 1980 ROBERT KRAMAR MARGIT PAVELKA Acknowledgements We wish to express our gratitude to Prof. Dr. D. H. FAHIMI and Dr. P. KALMBACH (Heidelberg) for kindly providing Figure 14, to Prof. Dr. KARIN GoRGAS (Heidelberg) for allowing the reproduction of Figure 43, to Profs. Dr. L. STOCKINGER and Dr. E. KAISER for helpful criticism, and to all our colleagues in our respective institutes. W e are especially grateful to Drs. H. GoLDENBERG and M. HuTTINGER for continuous discussion, to Mrs. JuTTA SELBMANN for typing the references, and to Mr. P. KAMPFER and Mr. H. WAGNER for carefully drawing some of the figures.
An invited collection of peer-reviewed papers surveying key areas of Roger Needham's distinguished research career at Cambridge University and Microsoft Research. From operating systems to distributed computing, many of the world's leading researchers provide insight into the latest concepts and theoretical insights--many of which are based upon Needham's pioneering research work. A critical collection of edited-survey research papers spanning the entire range of Roger Needham's distinguished scientific career, from operating systems to distributed computing and security. Many of the world's leading researchers survey their topics' latest developments and acknowledge the theoretical foundations of Needham's work. Introduction to book written by Rick Rashid, Director of Microsoft Research Worldwide.
JR is an extension of the Java programming language with additional concurrency mechanisms based on those in the SR (Synchronizing Resources) programming language. The JR implementation executes on UNIX-based systems (Linux, Mac OS X, and Solaris) and Windows-based systems. It is available free from the JR webpage. This book describes the JR programming language and illustrates how it can be used to write concurrent programs for a variety of applications. This text presents numerous small and large example programs. The source code for all programming examples and the given parts of all programming exercises are available on the JR webpage. Dr. Ronald A. Olsson and Dr. Aaron W. Keen, the authors of this text, are the designers and implementors of JR.
This book constitutes the refereed proceedings of the 26th International Conference on Architecture of Computing Systems, ARCS 2013, held in Prague, Czech Republic, in February 2013. The 29 papers presented were carefully reviewed and selected from 73 submissions. The topics covered are computer architecture topics such as multi-cores, memory systems, and parallel computing, adaptive system architectures such as reconfigurable systems in hardware and software, customization and application specific accelerators in heterogeneous architectures, organic and autonomic computing including both theoretical and practical results on self-organization, self-configuration, self-optimization, self-healing, and self-protection techniques, operating systems including but not limited to scheduling, memory management, power management, RTOS, energy-awareness, and green computing.
This book constitutes the proceedings of the 22nd International Conference on Compiler Construction, CC 2013, held as part of the European Joint Conferences on Theory and Practice of Software, ETAPS 2013, which took place in Rome, Italy, in March 2013. The 13 papers presented in this book were carefully reviewed and selected from 53 submissions. They have been organized into five topical sections on register allocation, pointer analysis, data and information flow, machine learning, and refactoring.
Real-Time Systems Engineering and Applications is a well-structured collection of chapters pertaining to present and future developments in real-time systems engineering. After an overview of real-time processing, theoretical foundations are presented. The book then introduces useful modeling concepts and tools. This is followed by concentration on the more practical aspects of real-time engineering with a thorough overview of the present state of the art, both in hardware and software, including related concepts in robotics. Examples are given of novel real-time applications which illustrate the present state of the art. The book concludes with a focus on future developments, giving direction for new research activities and an educational curriculum covering the subject. This book can be used as a source for academic and industrial researchers as well as a textbook for computing and engineering courses covering the topic of real-time systems engineering.
This book constitutes the proceedings of the 23rd International Conference on Compiler Construction, CC 2014, which was held as part of the European Joint Conferences on Theory and Practice of Software, ETAPS 2014, which took place in Grenoble, France, in April 2014. The 10 full papers and 4 tool papers included in this volume were carefully reviewed and selected from 47 submissions; the book also contains one invited talk. The papers are organized in topical sections named: program analysis and optimization; parallelism and parsing and new trends in compilation.
This book describes the key concepts, principles and implementation options for creating high-assurance cloud computing solutions. The guide starts with a broad technical overview and basic introduction to cloud computing, looking at the overall architecture of the cloud, client systems, the modern Internet and cloud computing data centers. It then delves into the core challenges of showing how reliability and fault-tolerance can be abstracted, how the resulting questions can be solved, and how the solutions can be leveraged to create a wide range of practical cloud applications. The author's style is practical, and the guide should be readily understandable without any special background. Concrete examples are often drawn from real-world settings to illustrate key insights. Appendices show how the most important reliability models can be formalized, describe the API of the Isis2 platform, and offer more than 80 problems at varying levels of difficulty.
This book constitutes the refereed proceedings of the Cryptographers' Track at the RSA Conference 2013, CT-RSA 2013, held in San Francisco, CA, USA, in February/March 2013. The 25 revised full papers presented were carefully reviewed and selected from 89 submissions. The papers are grouped into topical sections covering: side channel attacks, digital signatures, public-key encryption, cryptographic protocols, secure implementation methods, symmetric key primitives, and identity-based encryption.
This book constitutes the thoroughly refereed proceedings of the 16th International Workshop on Job Scheduling Strategies for Parallel Processing, JSSPP 2012, which was held in Shanghai, China, in May 2012. The 14 revised papers presented were carefully reviewed and selected from 24 submissions. The papers cover the following topics: parallel batch scheduling; workload analysis and modeling; resource management system software studies; and Web scheduling.
This book constitutes the refereed post-proceedings of the 9th IFIP International Conference on Network and Parallel Computing, NPC 2012, held in Gwangju, Korea, in September 2012. The 38 papers presented were carefully reviewed and selected from 136 submissions. The papers are organized in the following topical sections: algorithms, scheduling, analysis, and data mining; network architecture and protocol design; network security; paralel, distributed, and virtualization techniques; performance modeling, prediction, and tuning; resource management; ubiquitous communications and networks; and web, communication, and cloud computing. In addition, a total of 37 papers selected from five satellite workshops (ATIMCN, ATSME, Cloud&Grid, DATICS, and UMAS 2012) are included.
This book is a result of the Seventh International Conference on Information Sys tems Development-Methods and Tools, Theory and Practice held in Bled, Slovenia, Sep tember 21-23, 1998. The purpose of the conference was to address issues facing academia and industry when specifying, developing, managing, and improving information comput erized systems. During the past few years, many new concepts and approaches emerged in the Information Systems Development (ISD) field. The various theories, methods, and tools available to system developers also bring problems such as choosing the most effec tive approach for a specific task. This conference provides a meeting place for IS re searchers and practitioners from Eastern and Western Europe as well as from other parts of the world. An objective of the conference is not only to share scientific knowledge and in terests but to establish strong professional ties among the participants. The Seventh International Conference on Information Systems Develop ment-ISD'98 continues the concepts of the first Polish-Scandinavian Seminar on Current Trends in Information Systems Development Methodologies held in Gdansk, Poland in 1988. Through the years, the Seminar developed into the International Conference on In formation Systems Development. ISD'99 will be held in Boise, Idaho. The selection of papers was carried out by the International Program Committee. All papers were reviewed in advance by three people. Papers were judged according to their originality, relevance, and presentation quality. All papers were judged only on their own merits, independent of other submissions.
Real-time computing systems are vital to a wide range of applications. For example, they are used in the control of nuclear reactors and automated manufacturing facilities, in controlling and tracking air traffic, and in communication systems. In recent years, real-time systems have also grown larger and become more critical. For instance, advanced aircraft such as the space shuttle must depend heavily on computer sys tems Carlow 84]. The centralized control of manufacturing facilities and assembly plants operated by robots are other examples at the heart of which lie embedded real-time systems. Military defense systems deployed in the air, on the ocean surface, land and underwater, have also been increasingly relying upon real-time systems for monitoring and operational safety purposes, and for retaliatory and containment measures. In telecommunications and in multi-media applications, real time characteristics are essential to maintain the integrity of transmitted data, audio and video signals. Many of these systems control, monitor or perform critical operations, and must respond quickly to emergency events in a wide range of embedded applications. They are therefore required to process tasks with stringent timing requirements and must perform these tasks in a way that these timing requirements are guaranteed to be met. Real-time scheduling al gorithms attempt to ensure that system timing behavior meets its specifications, but typically assume that tasks do not share logical or physical resources. Since resource-sharing cannot be eliminated, synchronization primitives must be used to ensure that resource consis tency constraints are not violated."
Cooperating Heterogeneous Systems provides an in-depth introduction to the issues and techniques surrounding the integration and control of diverse and independent software components. Organizations increasingly rely upon diverse computer systems to perform a variety of knowledge-based tasks. This presents technical issues of interoperability and integration, as well as philosophical issues of how cooperation and interaction between computational entities is to be realized. Cooperating systems are systems that work together towards a common end. The concepts of cooperation must be realized in technically sound system architectures, having a uniform meta-layer between knowledge sources and the rest of the system. The layer consists of a family of interpreters, one for each knowledge source, and meta-knowledge. A system architecture to integrate and control diverse knowledge sources is presented. The architecture is based on the meta-level properties of the logic programming language Prolog. An implementation of the architecture is described, a Framework for Logic Programming Systems with Distributed Execution (FLiPSiDE). Knowledge-based systems play an important role in any up-to-date arsenal of decision support tools. The tremendous growth of computer communications infrastructure has made distributed computing a viable option, and often a necessity in geographically distributed organizations. It has become clear that to take knowledge-based systems to their next useful level, it is necessary to get independent knowledge-based systems to work together, much as we put together ad hoc work groups in our organizations to tackle complex problems. The book is for scientists and software engineers who have experience in knowledge-based systems and/or logic programming and seek a hands-on introduction to cooperating systems. Researchers investigating autonomous agents, distributed computation, and cooperating systems will find fresh ideas and new perspectives on well-established approaches to control, organization, and cooperation.
This book is a result of the ISD'99, Eight International Conference on Infonnation Systems Development-Methods and Tools, Theory, and Practice held August 11-13, 1999 in Boise, Idaho, USA. The purpose of this conference was to address the issues facing academia and industry when specifying, developing, managing, and improving infonnation systems. ISD'99 consisted not only of the technical program represented in these Proceedings, but also of plenary sessions on product support and content management systems for the Internet environment, workshop on a new paradigm for successful acquisition of infonnation systems, and a panel discussion on current pedagogical issues in systems analysis and design. The selection of papers for ISD'99 was carried out by the International Program Committee. Papers presented during the conference and printed in this volume have been selected from submissions after fonnal double-blind reviewing process and have been revised by their authors based on the recommendations of reviewers. Papers were judged according to their originality, relevance, and presentation quality. All papers were judged purely on their own merits, independently of other submissions. We would like to thank the authors of papers accepted for ISD'99 who all made gallant efforts to provide us with electronic copies of their manuscripts confonning to common guidelines. We thank them for thoughtfully responding to reviewers comments and carefully preparing their final contributions. We thank Daryl Jones, provost of Boise State University and William Lathen, dean, College of Business and Economics, for their support and encouragement.
This book constitutes the proceedings of the 10th IFIP International Conference on Network and Parallel Computing, NPC 2013, held in Guiyang, China, in September 2013. The 34 papers presented in this volume were carefully reviewed and selected from 109 submissions. They are organized in topical sections named: parallel programming and algorithms; cloud resource management; parallel architectures; multi-core computing and GPU; and miscellaneous.
Unique selling point: * This book proposes several approaches for dynamic Android malware detection based on system calls which do not have the limitations of existing mechanisms. * This book will be useful for researchers, students, developers and security analysts to know how malware behavior represented in the form of system call graphs can effectively detect Android malware. * The malware detection mechanisms in this book can be integrated with commercial antivirus softwares to detect Android malware including obfuscated variants.
In this international collection of papers there is a wealth of knowledge on artificial intelligence (AI) and cognitive science (CS) techniques applied to the problem of providing help systems mainly for the UNIX operating system. The research described here involves the representation of technical computer concepts, but also the representation of how users conceptualise such concepts. The collection looks at computational models and systems such as UC, Yucca, and OSCON programmed in languages such as Lisp, Prolog, OPS-5, and C which have been developed to provide UNIX help. These systems range from being menu-based to ones with natural language interfaces, some providing active help, intervening when they believe the user to have misconceptions, and some based on empirical studies of what users actually do while using UNIX. Further papers investigate planning and knowledge representation where the focus is on discovering what the user wants to do, and figuring out a way to do it, as well as representing the knowledge needed to do so. There is a significant focus on natural language dialogue where consultation systems can become active, incorporating user modfelling, natural language generation and plan recognition, modelling metaphors, and users' mistaken beliefs. Much can be learned from seeing how AI and CS techniques can be investigated in depth while being applied to a real test-bed domain such as help on UNIX.
An operating system is probably the most important part of the body of soft ware which goes with any modern computer system. I ts importance is reflected in the large amount of manpower usually invested in its construction, and in the mystique by which it is often surrounded. To the non-expert the design and construction of operating systems has often appeared an activity impenetrable to those who do not practise it. I hope this book will go some way toward dispelling the mystique, and encourage a greater general understanding of the principles on which operating systems are constructed. The material in the book is based on a course of lectures I have given for the past few years to undergraduate students of computer science. The book is therefore a suitable introduction to operating systems for students who have a basic grounding in computer science, or for people who have worked with computers for some time. Ideally the reader should have a knowledge of prorramming and be familiar with general machine architecture, common data structures such as lists and trees, and the functions of system software such as compilers, loaders, and editors. I t will also be helpful if he has had some experience of using a large operating system, seeing it, as it were, from the out side."
Compiler technology is fundamental to computer science since it provides the means to implement many other tools. It is interesting that, in fact, many tools have a compiler framework - they accept input in a particular format, perform some processing and present output in another format. Such tools support the abstraction process and are crucial to productive systems development. The focus of Compiler Technology: Tools, Translators and Language Implementation is to enable quick development of analysis tools. Both lexical scanner and parser generator tools are provided as supplements to this book, since a hands-on approach to experimentation with a toy implementation aids in understanding abstract topics such as parse-trees and parse conflicts. Furthermore, it is through hands-on exercises that one discovers the particular intricacies of language implementation. Compiler Technology: Tools, Translators and Language Implementation is suitable as a textbook for an undergraduate or graduate level course on compiler technology, and as a reference for researchers and practitioners interested in compilers and language implementation. |
You may like...
|