![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Computer hardware & operating systems > Operating systems & graphical user interfaces (GUIs)
The power of Core Data allows iOS developers to efficiently
store and retrieve application data using familiar object-oriented
paradigms.Pro Core Data for iOS explains both how and why to use
Core Data for data storage, from simple to advanced techniques.
Covering common and advanced persistence patterns, this book
prepares any iOS developer to store and retrieve data accurately
and proficiently. All iPhone, iPad, and iPod touch developers whose applications manage any amount of data. Table of Contents Getting Started Understanding Core Data Storing Data: SQLite and Other Options Creating a Data Model Working with Data Objects Refining Result Sets Tuning Performance and Memory Usage Versioning and Migrating Data Using Core Data in Advanced Applications
Want to learn how to program on your Mac? Not sure where to begin? Best-selling author Wallace Wang will explain how to get started with Cocoa, Objective-C, and Xcode. Whether you are an experienced Windows coder moving to the Mac, or you are completely new to programming, you'll see how the basic design of a Mac OS X program works, how Objective-C differs from other languages you may have used, and how to use the Xcode development environment. Most importantly, you'll learn how to use elements of the Cocoa framework to create windows, store data, and respond to users in your own Mac programs. If you want to learn how to develop apps with Cocoa, Objective-C, and Xcode, this book is a great first step. Here are just a few of the things you'll master along the way: Fundamental programming concepts aided by short, easy-to-understand examples How to use Xcode and related programming tools to save time and work more efficiently A firm understanding of the basics of Objective-C and how it compares to other languages you might know How to create simple apps using the Cocoa framework How to easily design, write, test, and market your finished program What you'll learn Basic programming concepts illustrated with short, easy-to-understand examples How to use Xcode and related programming tools The basics of Objective-C and how it compares to other languages you might know How to create simple apps using the Cocoa framework Who this book is for Mac users who want to learn programming or Windows developers who are moving to the Mac OS and iOS platforms. No previous programming experience required. Table of Contents Understanding Programming Understanding Apple's Programming Tools The Basic Steps to Creating a Mac Program Getting Help Learning Objective-C Making Decisions with Branches Repeating Code with Loops Understanding the Cocoa Framework Manipulating Strings Arrays Dictionaries and Sets Creating Classes and Objects Inheritance, Method Overriding, and Events Creating a User Interface Choosing Commands with Buttons Making Choices with Radio Buttons and Check Boxes Making Choices with Pop-Up Buttons Inputting and Outputting Data with Labels, Text Fields, and Combo Boxes Inputting Data with Sliders, Date Pickers, and Steppers Using Built-In Dialog Boxes Creating Pull-Down Menus Designing Your Own Programs Working with Xcode Debugging Your Program
iOS Forensic Analysis provides an in-depth look at investigative processes for the iPhone, iPod Touch, and iPad devices. The methods and procedures outlined in the book can be taken into any courtroom. With never-before-published iOS information and data sets that are new and evolving, this book gives the examiner and investigator the knowledge to complete a full device examination that will be credible and accepted in the forensic community. -->What you'll learn-->How to respond to security incidents involving iOS devices How to acquire and analyze data on iOS devices such as iPhone and iPad How to analyze media exploitation on iOS devices-->Who this book is for--> Computer forensic professionals, law enforcement, attorneys, security professionals, those who are curious about such things, and educators-->-->-->Table of Contents-->History of Apple Mobile Devices iOS Operating and File System Analysis Search, Seizure, and Incident Response iPhone Logical Acquisition Logical Data Analysis Mac and Windows Artifacts GPS Analysis Media Exploitation Media Exploitation Analysis Network Analysis
Mobile devices haveevolved to focus onrich media production and consumption. Developers of mobile applications are able to create applications that allow people to play, capture, andshare media in a variety of new ways on mobile devices. The popularity of Android has soared in part because the platform offers developers a rich set of capabilities including access to media capturing and playback functions. Pro Android Mediaprovides concise and clear instruction on how to utilize the media APIs made available through Android to create dynamic apps. It takes youfrom a simple means to gain access to the camera to complex video capture and sharing examples. It also covers sound, graphics, painting, and moreeverything you need to make your app come "alive." What you'll learn Develop graphics, music, video and rich media apps for Android smartphones and tablets Build touchscreen input features into Android apps that allow users to draw, paint, and do other creative forms of input Turn the Android smartphone into a full fledged media player Integrate and use location-based services and media-related web service APIs Who this book is for This book is aimed primarily at the growing market of Android developers. It is written in such a way that it may be used by those who are familiar with Android, but have no experience developing applications that deal with images, audio, or video Table of Contents Introduction to Android Imaging Building Custom Camera Applications Image Editing and Processing Graphics and Touch Events Introduction to Audio on Android Background and Networked Audio Audio Capture Audio Synthesis and Analysis Introduction to Video Advanced Video Video Capture Media Consumption and Publishing Using Web Services
Hard real-time systems are very predictable, but not sufficiently flexible to adapt to dynamic situations. They are built under pessimistic assumptions to cope with worst-case scenarios, so they often waste resources. Soft real-time systems are built to reduce resource consumption, tolerate overloads and adapt to system changes. They are also more suited to novel applications of real-time technology, such as multimedia systems, monitoring apparatuses, telecommunication networks, mobile robotics, virtual reality, and interactive computer games. This unique monograph provides concrete methods for building flexible, predictable soft real-time systems, in order to optimize resources and reduce costs. It is an invaluable reference for developers, as well as researchers and students in Computer Science.
With the advent of portable and autonomous computing systems, power con sumption has emerged as a focal point in many research projects, commercial systems and DoD platforms. One current research initiative, which drew much attention to this area, is the Power Aware Computing and Communications (PAC/C) program sponsored by DARPA. Many of the chapters in this book include results from work that have been supported by the PACIC program. The performance of computer systems has been tremendously improving while the size and weight of such systems has been constantly shrinking. The capacities of batteries relative to their sizes and weights has been also improv ing but at a rate which is much slower than the rate of improvement in computer performance and the rate of shrinking in computer sizes. The relation between the power consumption of a computer system and it performance and size is a complex one which is very much dependent on the specific system and the technology used to build that system. We do not need a complex argument, however, to be convinced that energy and power, which is the rate of energy consumption, are becoming critical components in computer systems in gen eral, and portable and autonomous systems, in particular. Most of the early research on power consumption in computer systems ad dressed the issue of minimizing power in a given platform, which usually translates into minimizing energy consumption, and thus, longer battery life."
Several works on multimedia storage appear in literature today, but very little if any, have been devoted to handling long duration video retrieval, over large scale networks. Distributed retrieval of multimedia documents, especially the long duration documents, is an imperative step in rendering high-quality, high-fidelity, and cost-effective services for network service providers. Distributed Multimedia Retrieval Strategies for Large Scale Networked Systems presents an up-to-date research status in the domain of distributed video retrieval. This professional book will include several different techniques that are in place for long duration video retrieval. An experimentally tested technology under the JINI platform, demonstrates a practical working system which serves as a feasibility study, as well as the first step in realizing such a technology.
Growing demands for the quality, safety, and security of software can only be satisfied by the rigorous application of formal methods during software design. This book methodically investigates the potential of first-order logic automated theorem provers for applications in software engineering. Illustrated by complete case studies on protocol verification, verification of security protocols, and logic-based software reuse, this book provides techniques for assessing the prover's capabilities and for selecting and developing an appropriate interface architecture.
This Festschrift volume, published in honor of Brian Randell on the occasion of his 75th birthday, contains a total of 37 refereed contributions. Two biographical papers are followed by the six invited papers that were presented at the conference 'Dependable and Historic Computing: The Randell Tales', held during April 7-8, 2011 at Newcastle University, UK. The remaining contributions are authored by former scientific colleagues of Brian Randell. The papers focus on the core of Brian Randell's work: the development of computing science and the study of its history. Moreover, his wider interests are reflected and so the collection comprises papers on software engineering, storage fragmentation, computer architecture, programming languages and dependability. There is even a paper that echoes Randell's love of maps. After an early career with English Electric and then with IBM in New York and California, Brian Randell joined Newcastle University. His main research has been on dependable computing in all its forms, especially reliability, safety and security aspects, and he has led several major European collaborative projects.
Lean Manufacturing has proved to be one of the most successful and most powerful production business systems over the last decades. Its application enabled many companies to make a big leap towards better utilization of resources and thus provide better service to the customers through faster response, higher quality and lowered costs. Lean is often described as "eyes for flow and eyes for muda" philosophy. It simply means that value is created only when all the resources flow through the system. If the flow is stopped no value but only costs and time are added, which is muda (Jap. waste). Since the philosophy was born at the Toyota many solutions were tailored for the high volume environment. But in turbulent, fast-changing market environment and progressing globalization, customers tend to require more customization, lower volumes and higher variety at much less cost and of better quality. This calls for adaptation of existing lean techniques and exploration of the new waste-free solutions that go far beyond manufacturing. This book brings together the opinions of a number of leading academics and researchers from around the world responding to those emerging needs. They tried to find answer to the question how to move forward from "Spaghetti World" of supply, production, distribution, sales, administration, product development, logistics, accounting, etc. Through individual chapters in this book authors present their views, approaches, concepts and developed tools. The reader will learn the key issues currently being addressed in production management research and practice throughout the world.
There is an established interest in integrating databases and programming languages. This book on Data Types and Persistence evolved from the proceedings of a workshop held at the Appin in August 1985. The purpose of the Appin workshop was to focus on these two aspects: persistence and data types, and to bring together people from various disciplines who have thought about these problems. Particular topics of"interest include the design of type systems appropriate for database work, the representation of persistent objects such as data types and modules, and the provision of orthogonal persistence and certain aspects of transactions and concurrency. The programme was broken into three sessions: morning, late afternoon and evening to allow the participants to take advantage of two beautiful days in the Scottish Highlands. The financial assistance of the Science and Engineering Research Council, the National Science Foundation and International Computers Ltd. is gratefully acknowledged. We would also like to thank Isabel Graham, Anne Donnelly and Estelle Taylor for their help in organising the workshop. Finally our thanks to Pete Bailey, Ray Carick and Dave Munro for the immense task they undertook in typesetting the book. The convergence of programming languages and databases to a coherent and consistent whole requires ideas from, and adjustment in, both intellectual camps. The first group of chapters in this book present ideas and adjustments coming from the programming language research community. This community frequently discusses types and uses them as a framework for other discussions.
The second half of the 1970s was marked with impressive advances in array/vector architectures and vectorization techniques and compilers. This progress continued with a particular focus on vector machines until the middle of the 1980s. The major ity of supercomputers during this period were register-to-register (Cray 1) or memory-to-memory (CDC Cyber 205) vector (pipelined) machines. However, the increasing demand for higher computational rates lead naturally to parallel comput ers and software. Through the replication of autonomous processors in a coordinated system, one can skip over performance barriers due technology limitations. In princi ple, parallelism offers unlimited performance potential. Nevertheless, it is very difficult to realize this performance potential in practice. So far, we have seen only the tip of the iceberg called "parallel machines and parallel programming." Parallel programming in particular is a rapidly evolving art and, at present, highly empirical. In this book we discuss several aspects of parallel programming and parallelizing compilers. Instead of trying to develop parallel programming methodologies and paradigms, we often focus on more advanced topics assuming that the reader has an adequate background in parallel processing. The book is organized in three main parts. In the first part (Chapters 1 and 2) we set the stage and focus on program transformations and parallelizing compilers. The second part of this book (Chapters 3 and 4) discusses scheduling for parallel machines from the practical point of view macro and microtasking and supporting environments). Finally, the last part (Le."
The term Wearable Technology encompasses a wide spectrum of devices, services and systems for wireless communications and the web. This book discusses characteristics and design elements required for wearable devices and systems to be embraced by the mainstream population for use in their everyday lives, introducing concepts such as Operational Inertia. The book discusses social and legal issues that may pose the greatest impediment to adoption of wearables. The book is structured to meet the needs of researchers and practitioners in industry, and can also be used as a secondary text in advanced-level courses in computer science and electrical engineering.
This book will attempt to give a first synthesis of recent works con cerning reactive system design. The term "reactive system" has been introduced in order to at'oid the ambiguities often associated with by the term "real-time system," which, although best known and more sugges tive, has been given so many different meanings that it is almost in evitably misunderstood. Industrial process control systems, transporta tion control and supervision systems, signal-processing systems, are ex amples of the systems we have in mind. Although these systems are more and more computerized, it is sur prising to notice that the problem of time in computer science has been studied only recently by "pure" computer scientists. Until the early 1980s, time problems were regarded as the concern of performance evalu ation, or of some (unjustly scorned) "industrial computer engineering," or, at best, of operating systems. A second surprising fact, in contrast, is the growth of research con cerning timed systems during the last decade. The handling of time has suddenly become a fundamental goal for most models of concurrency. In particular, Robin Alilner 's pioneering works about synchronous process algebras gave rise to a school of thought adopting the following abstract point of view: As soon as one admits that a system can instantaneously react to events, i. e."
This remarkable anthology allows the pioneers who orchestrated the major breakthroughs in operating system technology to describe their work in their own words. From the batch processing systems of the 1950s to the distributed systems of the 1990s, Tom Kilburn, David Howarth, Bill Lynch, Fernando Corbato, Robert Daley, Sandy Fraser, Dennis Ritchie, Ken Thompson, Edsger Dijkstra, Per Brinch Hansen, Soren Lauesen, Barbara Liskov, Joe Stoy, Christopher Strachey, Butler Lampson, David Redell, Brian Randell, Andrew Tanenbaum, and others describe the systems they designed. The volume details such classic operating systems as the Atlas, B5000, Exec II, Egdon, CTSS, Multics, Titan, Unix, THE, RC 4000, Venus, Boss 2, Solo, OS 6, Alto, Pilot, Star, WFS, Unix United, and Amoeba systems. An introductory essay on the evolution of operating systems summarizes the papers and helps puts them into a larger perspective. This provocative journey captures the historic contributions of operating systems to software design, concurrent programming, graphic user interfaces, file systems, personal computing, and distributed systems. It also fully portrays how operating systems designers think. It's ideal for everybody in the field, from students to professionals, academics to enthusiasts.
Real-Time Systems in Mechatronic Applications brings together in one place important contributions and up-to-date research results in this fast moving area. Real-Time Systems in Mechatronic Applications serves as an excellent reference, providing insight into some of the most challenging research issues in the field.
Security and Dependability for Ambient Intelligence is the primary publication of the SERENITY approach, which provides security and dependability (S&D) solutions for dynamic, highly distributed, heterogeneous systems. The objective of SERENITY is to enhance the security and dependability of ambient intelligence systems by providing a framework supporting the automated integration, configuration, monitoring and adaptation of security and dependability mechanisms. An edited volume contributed by world leaders in the field, this book covers the problems that the highly dynamic and heterogeneous nature of ambient intelligence systems poses to security and dependability and presents solutions to these problems. Security and Dependability for Ambient Intelligence is designed for researchers and practitioners focusing on the dynamic integration, deployment and verification of security and dependability solutions in highly distributed systems incorporating ambient intelligence features. It is also suitable as a reference or secondary text for advanced-level students in computer science and computer or electrical engineering.
The memory system is increasingly turning into a bottleneck in the design of embedded systems. The speed improvements of memory systems are lower than the speed improvements of processors, eventually leading to embedded systems whose performance is limited by the memory. This problem is known as the "memory wall" problem. Furthermore, memory systems may consume the largest share of the system s energy budget and may be the source of unpredictable timing behaviour. Hence, the design of the memory system deserves an increasing amount of attention. Fast, Efficient and Predictable Memory Accesses presents techniques for designing fast, energy-efficient and timing predictable memory systems. By using a careful combination of compiler optimizations and architectural improvements, we can achieve more than what would be feasible at one of the levels in isolation. The described optimization algorithms achieve the goals of high performance and low energy consumption. In addition to these benefits, the use of scratchpad memories significantly improves the timing predictability of the entire system, leading to tighter worst case execution time bounds (WCET). The WCET is a relevant design parameter for all timing critical systems. In addition, the book covers algorithms to exploit the power down modes of main memories in SDRAM technology, as well as the execute-in-place feature of Flash memories. The final chapter considers the impact of the register file, which is also part of the memory hierarchy."
Computer systems can only deliver benefits if functionality, users and usability are central to their design and deployment. This book encapsulates work done in the DIRC project (Interdisciplinary Research Collaboration in Dependability), bringing together a range of disciplinary approaches - computer science, sociology and software engineering - to produce a socio-technical systems perspective on the issues surrounding trust in technology in complex settings.
This book constitutes the refereed proceedings of Industry Oriented Conferences held at IFIP 20th World Computer Congress in September 2008. The IFIP series publishes state-of-the-art results in the sciences and technologies of information and communication. The scope of the series includes: foundations of computer science; software theory and practice; education; computer applications in technology; communication systems; systems modeling and optimization; information systems; computers and society; computer systems technology; security and protection in information processing systems; artificial intelligence; and human-computer interaction. Proceedings and post-proceedings of refereed international conferences in computer science and interdisciplinary fields are featured. These results often precede journal publication and represent the most current research. The principal aim of the IFIP series is to encourage education and the dissemination and exchange of information about all aspects of computing.
Storage Management in Data Centers helps administrators tackle the complexity of data center mass storage. It shows how to exploit the potential of Veritas Storage Foundation by conveying information about the design concepts of the software as well as its architectural background. Rather than merely showing how to use Storage Foundation, it explains why to use it in a particular way, along with what goes on inside. Chapters are split into three sections: An introductory part for the novice user, a full-featured part for the experienced, and a technical deep dive for the seasoned expert. An extensive troubleshooting section shows how to fix problems with volumes, plexes, disks and disk groups. A snapshot chapter gives detailed instructions on how to use the most advanced point-in-time copies. A tuning chapter will help you speed up and benchmark your volumes. And a special chapter on split data centers discusses latency issues as well as remote mirroring mechanisms and cross-site volume maintenance. All topics are covered with the technical know how gathered from an aggregate thirty years of experience in consulting and training in data centers all over the world.
Web caching and content delivery technologies provide the
infrastructure on which systems are built for the scalable
distribution of information. This proceedings of the eighth annual
workshop, captures a cross-section of the latest issues and
techniques of interest to network architects and researchers in
large-scale content delivery. Topics covered include the
distribution of streaming multimedia, edge caching and computation,
multicast, delivery of dynamic content, enterprise content
delivery, streaming proxies and servers, content transcoding,
replication and caching strategies, peer-to-peer content delivery,
and Web prefetching.
A detailed introduction to interdisciplinary application area of distributed systems, namely the computer support of individuals trying to solve a problem in cooperation with each other but not necessarily having identical work places or working times. The book is addressed to students of distributed systems, communications, information science and socio-organizational theory, as well as to users and developers of systems with group communication and cooperation as top priorities.
Based on research and industry experience, this book structures the issues pertaining to grid computing security into three main categories: architecture-related, infrastructure-related, and management-related issues. It discusses all three categories in detail, presents existing solutions, standards, and products, and pinpoints their shortcomings and open questions. Together with a brief introduction into grid computing in general and underlying security technologies, this book offers the first concise and detailed introduction to this important area, targeting professionals in the grid industry as well as students.
This year, the IFIP Working Conference on Distributed and Parallel Embedded Sys tems (DIPES 2008) is held as part of the IFIP World Computer Congress, held in Milan on September 7 10, 2008. The embedded systems world has a great deal of experience with parallel and distributed computing. Many embedded computing systems require the high performance that can be delivered by parallel computing. Parallel and distributed computing are often the only ways to deliver adequate real time performance at low power levels. This year's conference attracted 30 submissions, of which 21 were accepted. Prof. Jor ] g Henkel of the University of Karlsruhe graciously contributed a keynote address on embedded computing and reliability. We would like to thank all of the program committee members for their diligence. Wayne Wolf, Bernd Kleinjohann, and Lisa Kleinjohann Acknowledgements We would like to thank all people involved in the organization of the IFIP World Computer Congress 2008, especially the IPC Co Chairs Judith Bishop and Ivo De Lotto, the Organization Chair Giulio Occhini, as well as the Publications Chair John Impagliazzo. Further thanks go to the authors for their valuable contributions to DIPES 2008. Last but not least we would like to acknowledge the considerable amount of work and enthusiasm spent by our colleague Claudius Stern in preparing theproceedingsofDIPES2008. Hemadeitpossibletoproducethemintheircurrent professional and homogeneous style." |
![]() ![]() You may like...
An Introduction to XML and Web…
Anders Moller, Michael Schwartzbach
Paperback
R2,721
Discovery Miles 27 210
Privacy and Identity Management for Life…
Simone Fischer-Hubner, Penny Duquenoy, …
Hardcover
R1,573
Discovery Miles 15 730
|