![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer hardware & operating systems > Operating systems & graphical user interfaces (GUIs)
This volume contains a selection of papers that focus on the state-of the-art in real-time scheduling and resource management. Preliminary versions of these papers were presented at a workshop on the foundations of real-time computing sponsored by the Office of Naval Research in October, 1990 in Washington, D.C. A companion volume by the title Foundations of Real-Time Computing: Fonnal Specifications and Methods complements this book by addressing many of the most advanced approaches currently being investigated in the arena of formal specification and verification of real-time systems. Together, these two texts provide a comprehensive snapshot of current insights into the process of designing and building real-time computing systems on a scientific basis. Many of the papers in this book take care to define the notion of real-time system precisely, because it is often easy to misunderstand what is meant by that term. Different communities of researchers variously use the term real-time to refer to either very fast computing, or immediate on-line data acquisition, or deadline-driven computing. This text is concerned with the very difficult problems of scheduling tasks and resource management in computer systems whose performance is inextricably fused with the achievement of deadlines. Such systems have been enabled for a rapidly increasing set of diverse end-uses by the unremitting advances in computing power per constant-dollar cost and per constant-unit-volume of space. End-use applications of deadline-driven real-time computers span a spectrum that includes transportation systems, robotics and manufacturing, aerospace and defense, industrial process control, and telecommunications."
As we continue to build faster and fast. er computers, their performance is be coming increasingly dependent on the memory hierarchy. Both the clock speed of the machine and its throughput per clock depend heavily on the memory hierarchy. The time to complet. e a cache acce88 is oft. en the factor that det. er mines the cycle time. The effectiveness of the hierarchy in keeping the average cost of a reference down has a major impact on how close the sustained per formance is to the peak performance. Small changes in the performance of the memory hierarchy cause large changes in overall system performance. The strong growth of ruse machines, whose performance is more tightly coupled to the memory hierarchy, has created increasing demand for high performance memory systems. This trend is likely to accelerate: the improvements in main memory performance will be small compared to the improvements in processor performance. This difference will lead to an increasing gap between prOCe880r cycle time and main memory acce. time. This gap must be closed by improving the memory hierarchy. Computer architects have attacked this gap by designing machines with cache sizes an order of magnitude larger than those appearing five years ago. Microproce880r-based RISe systems now have caches that rival the size of those in mainframes and supercomputers."
This book is on dependence concepts and general methods for dependence testing. Here, dependence means data dependence and the tests are compile-time tests. We felt the time was ripe to create a solid theory of the subject, to provide the research community with a uniform conceptual framework in which things fit together nicely. How successful we have been in meeting these goals, of course, remains to be seen. We do not try to include all the minute details that are known, nor do we deal with clever tricks that all good programmers would want to use. We do try to convince the reader that there is a mathematical basis consisting of theories of bounds of linear functions and linear diophantine equations, that levels and direction vectors are concepts that arise rather natu rally, that different dependence tests are really special cases of some general tests, and so on. Some mathematical maturity is needed for a good understand ing of the book: mainly calculus and linear algebra. We have cov ered diophantine equations rather thoroughly and given a descrip of some matrix theory ideas that are not very widely known. tion A reader familiar with linear programming would quickly recog nize several concepts. We have learned a great deal from the works of M. Wolfe, and K. Kennedy and R. Allen. Wolfe's Ph. D. thesis at the University of Illinois and Kennedy & Allen's paper on vectorization of Fortran programs are still very useful sources on this subject."
Smart Home Automation with Linux and Raspberry Pi shows you how to automate your lights, curtains, music, and more, and control everything via a laptop or mobile phone. You'll learn how to use Linux, including Linux on Raspberry Pi, to control appliances and everything from kettles to curtains, including how to hack game consoles and even incorporate LEGO Mindstorms into your smart home schemes. You'll discover the practicalities on wiring a house in terms of both and power and networking, along with the selection and placement of servers. There are also explanations on handling communication to (and from) your computer with speech, SMS, email, and web. Finally, you'll see how your automated appliances can collaborate to become a smart home. Smart Home Automation with Linux was already an excellent resource for home automation, and in this second edition, Steven Goodwin will show you how a house can be fully controlled by its occupants, all using open source software and even open source hardware like Raspberry Pi and Arduino. What you'll learn Control appliances like kettles and curtains both locally and remotely. Find and harness data sources to provide context-aware living. Hack/change existing hardware/software to better fit your needs. Integrate various technologies into a function greater than the whole. Set up a home network, for both network and audio/video traffic. Learn how to incorporate Raspberry Pi, Arduino, and even LEGO Mindstorms into your smart home. Who this book is for This book is for amateur and professional Linux users and electronics enthusiasts who want to control their homes and their gadgets.
OS X and iOS Kernel Programming combines essential operating system and kernel architecture knowledge with a highly practical approach that will help you write effective kernel-level code. You'll learn fundamental concepts such as memory management and thread synchronization, as well as the I/O Kit framework. You'll also learn how to write your own kernel-level extensions, such as device drivers for USB and Thunderbolt devices, including networking, storage and audio drivers. OS X and iOS Kernel Programming provides an incisive and complete introduction to the XNU kernel, which runs iPhones, iPads, iPods, and Mac OS X servers and clients. Then, you'll expand your horizons to examine Mac OS X and iOS system architecture. Understanding Apple's operating systems will allow you to write efficient device drivers, such as those covered in the book, using I/O Kit. With OS X and iOS Kernel Programming, you'll: Discover classical kernel architecture topics such as memory management and thread synchronization Become well-versed in the intricacies of the kernel development process by applying kernel debugging and profiling tools Learn how to deploy your kernel-level projects and how to successfully package them Write code that interacts with hardware devices Examine easy to understand example code that can also be used in your own projects Create network filters Whether you're a hobbyist, student, or professional engineer, turn to OS X andiOS Kernel Programming and find the knowledge you need to start developing What you'll learn OS X and iOS common core architecture How to write extremely efficient code by exploiting kernel details Coding kernel-level extensions How to write device drivers How to program the I/O Kit framework Key mobile device topics like power management driversand video capture modules To understand OS X memory management and threads To parse kernel debug messages and package projects ready for deployment Who this book is for This book is suited for Intermediate and advanced iPhone and OS X programmers ready for the next step Kernel-level programmers interested in how OS X and iOS function Open source programmers with a background in Linuxor BSD, OS X and iOS Programmers interested in application performance System administrators running OS X clusters Table of Contents Operating System Fundamentals Mac OS X and iOS Xcode and the Kernel Development Environment The I/O Kit Framework Interacting with Drivers fromApplications Memory Management Synchronisation and Threading USB Drivers PCI and Thunderbolt Power Management Serial Port Drivers Core Audio Network Drivers Storage Drivers and Filesystems User-Space Drivers Debugging and Profiling Advanced Kernel Programming Deployment
This book introduces the fundamental concepts and practical simulation te- niques for modeling different aspects of operating systems to study their g- eral behavior and their performance. The approaches applied are obje- oriented modeling and process interaction approach to discrete-event simu- tion. The book depends on the basic modeling concepts and is more specialized than my previous book: Practical Process Simulation with Object-Oriented Techniques and C++, published by Artech House, Boston 1999. For a more detailed description see the Web location: http: //science.kennesaw.edu/ jgarrido/mybook, html. Most other books on performance modeling use only analytical approaches, and very few apply these concepts to the study of operating systems. Thus, the unique feature of the book is that it concentrates on design aspects of operating systems using practical simulation techniques. In addition, the book illustrates the dynamic behavior of different aspects of operating systems using the various simulation models, with a general hands-on approac
Parallel Language and Compiler Research in Japan offers the international community an opportunity to learn in-depth about key Japanese research efforts in the particular software domains of parallel programming and parallelizing compilers. These are important topics that strongly bear on the effectiveness and affordability of high performance computing systems. The chapters of this book convey a comprehensive and current depiction of leading edge research efforts in Japan that focus on parallel software design, development, and optimization that could be obtained only through direct and personal interaction with the researchers themselves.
This book constitutes the refereed proceedings of the 14th International Conference on Passive and Active Measurement, PAM 2013, held in Hong Kong, China, in March 2013. The 24 revised full papers presented were carefully reviewed and selected from 74 submissions. The papers have been organized in the following topical sections: measurement design, experience and analysis; Internet wireless and mobility; performance measurement; protocol and application behavior; characterization of network usage; and network security and privacy. In addition, 9 poster abstracts have been included.
The first ESPRIT Basic Research Project on Predictably Dependable Computing Systems (No. 3092, PDCS) commenced in May 1989, and ran until March 1992. The institutions and principal investigators that were involved in PDCS were: City University, London, UK (Bev Littlewood), lEI del CNR, Pisa, Italy (Lorenzo Strigini), Universitiit Karlsruhe, Germany (Tom Beth), LAAS-CNRS, Toulouse, France (Jean-Claude Laprie), University of Newcastle upon Tyne, UK (Brian Randell), LRI-CNRS/Universite Paris-Sud, France (Marie-Claude Gaudel), Technische Universitiit Wien, Austria (Hermann Kopetz), and University of York, UK (John McDermid). The work continued after March 1992, and a three-year successor project (No. 6362, PDCS2) officially started in August 1992, with a slightly changed membership: Chalmers University of Technology, Goteborg, Sweden (Erland Jonsson), City University, London, UK (Bev Littlewood), CNR, Pisa, Italy (Lorenzo Strigini), LAAS-CNRS, Toulouse, France (Jean-Claude Laprie), Universite Catholique de Louvain, Belgium (Pierre-Jacques Courtois), University of Newcastle upon Tyne, UK (Brian Randell), LRI-CNRS/Universite Paris-Sud, France (Marie-Claude Gaudel), Technische Universitiit Wien, Austria (Hermann Kopetz), and University of York, UK (John McDermid). The summary objective of both projects has been "to contribute to making the process of designing and constructing dependable computing systems much more predictable and cost-effective." In the case of PDCS2, the concentration has been on the problems of producing dependable distributed real-time systems and especially those where the dependability requirements centre on issues of safety and/or security.
This book constitutes the refereed proceedings of the 6th International Conference on Provable Security, ProvSec 2012, held in Chengdu, China, in September 2012. The 16 full papers and 4 short papers presented were carefully reviewed and selected from 66 submissions. The papers are grouped in topical sections on signature schemes, foundations, leakage resilence and key escrow, encryption schemes, and information theoretical security.
This book constitutes the thoroughly refereed post-conference
proceedings of the 6th International Workshop on Security and Trust
Management, STM 2010, held in Athens, Greece, in September 2010.
This volume presents the refereed proceedings of the 7th International ICST Conference on Security and Privacy in Communication Networks, SecureComm 2011, held in London, UK, in September 2011. The 35 revised papers included in the volume were carefully reviewed and selected from numerous submissions. Topics covered include network intrusion detection; anonymity and privacy; wireless security; system security; DNS and routing security; and key management.
This book constitutes the refereed proceedings of the 9th International Symposium on Advanced Parallel Processing Technologies, APPT 2011, held in Shanghai, China, in September 2011. The 13 revised full papers presented were carefully reviewed and selected from 40 submissions. The papers are organized in topical sections on parallel distributed system architectures, architecture, parallel application and software, distributed and cloud computing.
PC viruses are not necessarily a major disaster despite what is sometimes written about them. But a virus infection is at the very least a nuisance, and potentially can lead to loss of data. Quite often it is the user's panic reaction to discovering a virus infection that does more than the virus itself. This book demystifies PC viruses, providing clear, accurate information about this relatively new PC problem. It enables managers and PC users to formulate an appropriate response; adequate for prevention and cure, but not `over the top'. Over 100 PC viruses and variants are documented in detail. You are told how to recognise each one, what it does, how it copies itself, and how to get rid of it. Other useful and relevant technical information is also provided. Strategies for dealing with potential and actual virus outbreaks are described for business, academic and other environments, with the emphasis on sensible but not unreasonable precautions. All users of IBM PC or compatible computers - from single machines to major LAN's - will find this book invaluable. All that is required is a working knowledge of DOS. Dr. Alan Solomon has been conducting primary research into PC viruses since they first appeared, and has developed the best-selling virus protection software Dr. Solomon's Anti-Virus Toolkit.
This book constitutes the refereed proceedings of the 12th IFIP TC 6/TC 11 International Conference on Communications and Multimedia Security, CMS 2010, held in Ghent, Belgium, in October 2011. The 26 revised papers presented were carefully reviewed and selected from 52 submissions. The papers are organized in topical sections on usability, architecture and framework security, mobile identity management, secure hardware platforms, biometrics, multimedia security, network security and authentication.
So how do you build an application for the iPhone and iPad? Don't you need to spend years learning complicated programming languages? What about Objective-C and Cocoa touch? The answer is that you don't need to know any of those things Anybody can start building simple apps for the iPhone and iPad, and this book will show you how. This update of an Apress bestseller walks you through creating your first app, using plain English and practical examples using the iOS 6 software development platform and more. It cuts through the fog of jargon and misinformation that surrounds iPhone and iPad app development, and gives you simple, step-by-step instructions to get you started. * Teaches iPhone and iPad apps development in language anyone can understand * Provides simple, step-by-step examples that make learning easy, using iOS 6 * Offers bonus videos from the author that enable you to follow along--it's like your own private classroom The iPhone is the hottest gadget of our generation, and much of its success has been fueled by the App Store, Apple's online marketplace for iPhone apps.Over 1 billion apps were downloaded during the nine months following the launch of the App Store, ranging from the simplest games to the most complex business apps. Everyone has an idea for the next bestselling iPhone app--that's why you're reading this now And with the popularity of the iPad, this demand will just continue to grow. What you'll learn * Get both yourself and your computer set up for iPhone and iPad application development * Start by making small changes to existing applications to build your knowledge and experience before creating your own applications * Follow steps in plain English to build simple apps and get them working immediately * Style your application so that it looks good and users can easily navigate through it * Make use of the iPhone's touch screen and accelerometer * Use shortcuts and cheat sheets to create apps the easy way Who this book is for If you have a great idea for an iPhone or iPad app, but have never programmed before, then this book is for you. You don't need to have any previous computer programming skills--as long as you have a desire to learn, and you know which end of the mouse is which, you'll be fine.
The two-volume set LNCS 6852/6853 constitutes the refereed proceedings of the 17th International Euro-Par Conference held in Bordeaux, France, in August/September 2011.The 81 revised full papers presented were carefully reviewed and selected from 271 submissions. The papers are organized in topical sections on support tools and environments; performance prediction and evaluation; scheduling and load-balancing; high-performance architectures and compilers; parallel and distributed data management; grid, cluster and cloud computing; peer to peer computing; distributed systems and algorithms; parallel and distributed programming; parallel numerical algorithms; multicore and manycore programming; theory and algorithms for parallel computation; high performance networks and mobile ubiquitous computing.
It is universally accepted today that parallel processing is here to stay but that software for parallel machines is still difficult to develop. However, there is little recognition of the fact that changes in processor architecture can significantly ease the development of software. In the seventies the availability of processors that could address a large name space directly, eliminated the problem of name management at one level and paved the way for the routine development of large programs. Similarly, today, processor architectures that can facilitate cheap synchronization and provide a global address space can simplify compiler development for parallel machines. If the cost of synchronization remains high, the pro gramming of parallel machines will remain significantly less abstract than programming sequential machines. In this monograph Bob Iannucci presents the design and analysis of an architecture that can be a better building block for parallel machines than any von Neumann processor. There is another very interesting motivation behind this work. It is rooted in the long and venerable history of dataflow graphs as a formalism for ex pressing parallel computation. The field has bloomed since 1974, when Dennis and Misunas proposed a truly novel architecture using dataflow graphs as the parallel machine language. The novelty and elegance of dataflow architectures has, however, also kept us from asking the real question: "What can dataflow architectures buy us that von Neumann ar chitectures can't?" In the following I explain in a round about way how Bob and I arrived at this question."
This book constitutes the refereed proceedings of the 6th IFIP WG 6.6 International Conference on Autonomous Infrastructure, Management, and Security, AIMS 2012, held in Luxembourg in June 2012. The 10 full papers presented were carefully reviewed and selected from 23 submissions. They cover autonomic and distributed management, network security, network monitoring, and special environments and Internet of Things. In addition, this book contains 9 workshop papers which were selected from 18 submissions. They deal with high-speed networks and network management, intrusion detection, and network monitoring and security.
This book constitutes the refereed proceedings of the 16th Australasian Conference on Information Security and Privacy, ACISP 2011, held in Melbourne, Australia, in July 2011. The 24 revised full papers presented together with an invited talk and 9 poster papers were carefully reviewed and selected from 103 submissions. The papers are organized in topical sections on symmetric key cryptography, hash functions, cryptographic protocols, access control and security, and public key cryptography.
The two-volume set LNCS 6852/6853 constitutes the refereed proceedings of the 17th International Euro-Par Conference held in Bordeaux, France, in August/September 2011. The 81 revised full papers presented were carefully reviewed and selected from 271 submissions. The papers are organized in topical sections on support tools and environments; performance prediction and evaluation; scheduling and load-balancing; high-performance architectures and compilers; parallel and distributed data management; grid, cluster and cloud computing; peer to peer computing; distributed systems and algorithms; parallel and distributed programming; parallel numerical algorithms; multicore and manycore programming; theory and algorithms for parallel computation; high performance networks and mobile ubiquitous computing.
This book constitutes the thoroughly refereed post-proceedings of the 16th International Workshop on Security Protocols, SP 2008, held in Cambridge, UK, in April 2008. The 17 revised full papers presented together with edited transcriptions of some of the discussions following the presentations have gone through multiple rounds of reviewing, revision, and selection. The theme of this workshop was "Remodelling the Attacker" with the intention to tell the students at the start of a security course that it is very important to model the attacker, but like most advice to the young, this is an oversimplification. Shouldn't the attacker's capability be an output of the design process as well as an input? The papers and discussions in this volume examine the theme from the standpoint of various different applications and adversaries.
If you've ever wondered how Linux carries out the complicated tasks assigned to it by the IP protocols -- or if you just want to learn about modern networking through real-life examples -- Understanding Linux Network Internals is for you. Like the popular O'Reilly book, Understanding the Linux Kernel, this book clearly explains the underlying concepts and teaches you how to follow the actual C code that implements it. Although some background in the TCP/IP protocols is helpful, you can learn a great deal from this text about the protocols themselves and their uses. And if you already have a base knowledge of C, you can use the book's code walkthroughs to figure out exactly what this sophisticated part of the Linux kernel is doing. Part of the difficulty in understanding networks -- and implementing them -- is that the tasks are broken up and performed at many different times by different pieces of code. One of the strengths of this book is to integrate the pieces and reveal the relationships between far-flung functions and data structures. Understanding Linux Network Internals is both a big-picture discussion and a no-nonsense guide to the details of Linux networking. Topics include: * Key problems with networking * Network interface card (NIC) device drivers * System initialization * Layer 2 (link-layer) tasks and implementation * Layer 3 (IPv4) tasks and implementation * Neighbor infrastructure and protocols (ARP) * Bridging * Routing * ICMP Author Christian Benvenuti, an operating system designer specializing in networking, explains much more than how Linux code works. He shows the purposes of major networking features and the trade-offs involved in choosing one solution over another. A large number of flowcharts and other diagrams enhance the book's understandability.
This book constitutes the refereed proceedings of the 6th International Conference on Mathematical Methods, Models, and Architectures for Computer Network Security, MMM-ACNS 2012, held in St. Petersburg, Russia in October 2012. The 14 revised full papers and 8 revised short presentations were carefully reviewed and selected from a total of 44 submissions. The papers are organized in topical sections on applied cryptography and security protocols, access control and information protection, security policies, security event and information management, instrusion prevention, detection and response, anti-malware techniques, security modeling and cloud security.
This volume contains a selection of papers that focus on the state-of the-art in formal specification and verification of real-time computing systems. Preliminary versions of these papers were presented at a workshop on the foundations of real-time computing sponsored by the Office of Naval Research in October, 1990 in Washington, D. C. A companion volume by the title Foundations of Real-Time Computing: Scheduling and Resource Management complements this hook by addressing many of the recently devised techniques and approaches for scheduling tasks and managing resources in real-time systems. Together, these two texts provide a comprehensive snapshot of current insights into the process of designing and building real time computing systems on a scientific basis. The notion of real-time system has alternative interpretations, not all of which are intended usages in this collection of papers. Different communities of researchers variously use the term real-time to refer to either very fast computing, or immediate on-line data acquisition, or deadline-driven computing. This text is concerned with the formal specification and verification of computer software and systems whose correct performance is dependent on carefully orchestrated interactions with time, e. g., meeting deadlines and synchronizing with clocks. Such systems have been enabled for a rapidly increasing set of diverse end-uses by the unremitting advances in computing power per constant-dollar cost and per constant-unit-volume of space. End use applications of real-time computers span a spectrum that includes transportation systems, robotics and manufacturing, aerospace and defense, industrial process control, and telecommunications." |
You may like...
Governing Human Lives and Health in…
Matilda Hellman, Tom Kettunen, …
Hardcover
R4,503
Discovery Miles 45 030
Australian Vietnamese Golf Association…
Tri Tue Tran, Hien Minh Thi Tran, …
Hardcover
R935
Discovery Miles 9 350
|