![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming
This book teaches the basics of XML with an original approach, using real-world examples from an interesting (and operating) environment with broad applicability. It covers the full spectrum of Berkeley DB XML tools, including the command-line shell, transactions, rollbacks, replication, archiving and monitoring. Techniques and concepts that have broad applicability outside of the subject matter are skillfully explained: XML, XPath, XQuery, XML schemas, all industry-standard technologies that find one of their best tutorial treatments, and all in the context of a simple database solution. The book also presents a remarkable example of query power.
In the realm of CAD & Office Integration a new technology has been introduced that will overturn many accepted ideas, both for developer and end-user. What is this revolutionary new technology? Called "OLE for Design and Modeling" it is an enhancement of Microsoft Windows OLE for high performance CAD/CAM/CAE-software. With this book, the reader will understand, how OLE for D&M enables users to introduce and manipulate CAD models within regular text-processing and DTP-documents. And why even high-end 3D design objects can now be transferred between different systems using easy "drag and drop" operations. Furthermore this "plug and play" CAD technology makes it possible to use older CAD documents and older software in an entirely new context. If you want to know, where technology is going to, you should read it.
In "Distributed Algorithms," Nancy Lynch provides a blueprint
for designing, implementing, and analyzing distributed algorithms.
She directs her book at a wide audience, including students,
programmers, system designers, and researchers. "Distributed Algorithms" contains the most significant
algorithms and impossibility results in the area, all in a simple
automata-theoretic setting. The algorithms are proved correct, and
their complexity is analyzed according to precisely defined
complexity measures. The problems covered include resource
allocation, communication, consensus among distributed processes,
data consistency, deadlock detection, leader election, global
snapshots, and many others. The material is organized according to the system model first by
the timing model and then by the interprocess communication
mechanism. The material on system models is isolated in separate
chapters for easy reference. The presentation is completely rigorous, yet is intuitive enough for immediate comprehension. This book familiarizes readers with important problems, algorithms, and impossibility results in the area: readers can then recognize the problems when they arise in practice, apply the algorithms to solve them, and use the impossibility results to determine whether problems are unsolvable. The book also provides readers with the basic mathematical tools for designing new algorithms and proving new impossibility results. In addition, it teaches readers how to reason carefully about distributed algorithms to model them formally, devise precise specifications for their required behavior, prove their correctness, and evaluate their performance with realistic measures."
In this book, three main notions will be used in the editors search of improvements in various areas of computer graphics: Artificial Intelligence, Viewpoint Complexity and Human Intelligence. Several Artificial Intelligence techniques are used in presented intelligent scene modelers, mainly declarative ones. Among them, the mostly used techniques are Expert systems, Constraint Satisfaction Problem resolution and Machine-learning. The notion of viewpoint complexity, that is complexity of a scene seen from a given viewpoint, will be used in improvement proposals for a lot of computer graphics problems like scene understanding, virtual world exploration, image-based modeling and rendering, ray tracing and radiosity. Very often, viewpoint complexity is used in conjunction with Artificial Intelligence techniques like Heuristic search and Problem resolution. The notions of artificial Intelligence and Viewpoint Complexity may help to automatically resolve a big number of computer graphics problems. However, there are special situations where is required to find a particular solution for each situation. In such a case, human intelligence has to replace, or to be combined with, artificial intelligence. Such cases, and proposed solutions are also presented in this book.
Designing Sorting Networks: A New Paradigm provides an in-depth guide to maximizing the efficiency of sorting networks, and uses 0/1 cases, partially ordered sets and Haase diagrams to closely analyze their behavior in an easy, intuitive manner. This book also outlines new ideas and techniques for designing faster sorting networks using Sortnet, and illustrates how these techniques were used to design faster 12-key and 18-key sorting networks through a series of case studies. Finally, it examines and explains the mysterious behavior exhibited by the fastest-known 9-step 16-key network. Designing Sorting Networks: A New Paradigm is intended for advanced-level students, researchers and practitioners as a reference book. Academics in the fields of computer science, engineering and mathematics will also find this book invaluable.
This 2nd edition textbook has been expanded to include of 175 additional pages of additional content, created in response to readers feedback, as well as to new hardware and software releases. The book presents foundational robotics concepts using the ROBOTIS BIOLOID and OpenCM-904 robotic systems, and is suitable as a curriculum for a first course in robotics for undergraduate students or a self-learner. It covers wheel-based robots, as well as walking robots. Although it uses the standard "Sense, Think, Act" approach, communications (bot-to-bot and PC-to-bot) programming concepts are treated in more depth (wired and wireless ZigBee/BlueTooth). Algorithms are developed and described via ROBOTIS' proprietary RoboPlus IDE, as well as the more open Arduino-based Embedded C environments. Additionally, a vast array of web-based multimedia materials are used for illustrating robotics concepts, code implementations and videos of actual resulting robot behaviors. Advanced sensor interfacing for gyroscope, inertial measuring unit, foot pressure sensor and color camera are also demonstrated.
Aimed at improving a programmers ability for altering code to fit changing requirements and for detecting and correcting errors, this book argues for a new way of thinking about maintaining software. It proposes the use of a set of human factors principles that govern the programmer-software-event world interactions and form the core of the maintenance process. The book is thus highly valuable for systems analysts and programmers, managers seeking to reduce costs, researchers looking at solutions to the maintenance problem, and students learning to write clear unambiguous programs.
It is well known that embedded systems have to be implemented efficiently. This requires that processors optimized for certain application domains are used in embedded systems. Such an optimization requires a careful exploration of the design space, including a detailed study of cost/performance tradeoffs. In order to avoid time-consuming assembly language programming during design space exploration, compilers are needed. In order to analyze the effect of various software or hardware configurations on the performance, retargetable compilers are needed that can generate code for numerous different potential hardware configurations. This book provides a comprehensive and up-to-date overview of the fast developing area of retargetable compilers for embedded systems. It describes a large set important tools as well as applications of retargetable compilers at different levels in the design flow. Retargetable Compiler Technology for Embedded Systems is mostly self-contained and requires only fundamental knowledge in software and compiler design. It is intended to be a key reference for researchers and designers working on software, compilers, and processor optimization for embedded systems.
This book contains extended and revised versions of the best papers that were presented during the fifteenth edition of the IFIP/IEEE WG10.5 International Conference on Very Large Scale Integration, a global System-on-a-Chip Design & CAD conference. The 15th conference was held at the Georgia Institute of Technology, Atlanta, USA (October 15-17, 2007). Previous conferences have taken place in Edinburgh, Trondheim, Vancouver, Munich, Grenoble, Tokyo, Gramado, Lisbon, Montpellier, Darmstadt, Perth and Nice. The purpose of this conference, sponsored by IFIP TC 10 Working Group 10.5 and by the IEEE Council on Electronic Design Automation (CEDA), is to provide a forum to exchange ideas and show industrial and academic research results in the field of microelectronics design. The current trend toward increasing chip integration and technology process advancements brings about stimulating new challenges both at the physical and system-design levels, as well in the test of these systems. VLSI-SoC conferences aim to address these exciting new issues.
The book gives a thorough introduction into object orientated design and programming using C++. At the same time it can be used as a library of very useful programs chosen from the fields of finance, adminstration and statistics. These include programs for calculating loan periods, amortization, least squares fitting, a spelling checker, Gregorian calendar, data compression and encryption, searching and sorting. Basic C++ programming is introduced with simple introductory programs while object-oriented programming in C++ is explained as we develop useful classes. Finally we give an introduction into object orientated design and we demonstrate its power by developing a banking package.
What will business software look like in the future? And how will it be developed? This book covers the proceedings of the first international conference on Future Business Software - a new think tank discussing the trends in enterprise software with speakers from Europe's most successful software companies and the leading research institutions. The articles focus on two of the most prominent trends in the field: emergent software and agile development processes. "Emergent Software" is a new paradigm of software development that addresses the highly complex requirements of tomorrow's business software and aims at dynamically and flexibly combining a business software solution's different components in order to fulfill customers' needs with a minimum of effort. Agile development processes are the response of software technology to the implementation of diverse and rapidly changing software requirements. A major focus is on the minimization of project risks, e.g. through short, iterative development cycles, test-driven development and an intensive culture of communication."
Generic programming is about making programs more adaptable by making them more general. Generic programs often embody non-traditional kinds of polymorphism; ordinary programs are obtained from them by suitably instantiating their parameters. In contrast with normal programs, the parameters of a generic program are often quite rich in structure; for example, they may be other programs, types or type constructors, class hierarchies, or even programming paradigms. Generic programming techniques have always been of interest, both to practitioners and to theoreticians, but only recently have generic programming techniques become a specific focus of research in the functional and object-oriented programming language communities. Generic Programming comprises the edited proceedings of the Working Conference on Generic Programming, which was sponsored by the International Federation for Information Processing (IFIP) and held in Dagstuhl, Germany in July 2002. With contributions from leading researchers around the world, this volume captures the state of the art in this important emerging area.
Software architectures have gained wide popularity in the last decade. They generally play a fundamental role in coping with the inherent difficulties of the development of large-scale and complex software systems. Component-oriented and aspect-oriented programming enables software engineers to implement complex applications from a set of pre-defined components. Software Architectures and Component Technology collects excellent chapters on software architectures and component technologies from well-known authors, who not only explain the advantages, but also present the shortcomings of the current approaches while introducing novel solutions to overcome the shortcomings. The unique features of this book are: evaluates the current architecture design methods and component composition techniques and explains their shortcomings; presents three practical architecture design methods in detail; gives four industrial architecture design examples; presents conceptual models for distributed message-based architectures; explains techniques for refining architectures into components; presents the recent developments in component and aspect-oriented techniques; explains the status of research on Piccola, Hyper/JA(R), Pluggable Composite Adapters and Composition Filters. Software Architectures and Component Technology is a suitable text for graduate level students in computer science and engineering, and as a reference for researchers and practitioners in industry.
For more and more systems, software has moved from a peripheral to a central role, replacing mechanical parts and hardware and giving the product a competitive edge. Consequences of this trend are an increase in: the size of software systems, the variability in software artifacts, and the importance of software in achieving the system-level properties. Software architecture provides the necessary abstractions for managing the resulting complexity. We here introduce the Third Working IEEFlIFIP Conference on Software Architecture, WICSA3. That it is already the third such conference is in itself a clear indication that software architecture continues to be an important topic in industrial software development and in software engineering research. However, becoming an established field does not mean that software architecture provides less opportunity for innovation and new directions. On the contrary, one can identify a number of interesting trends within software architecture research. The first trend is that the role of the software architecture in all phases of software development is more explicitly recognized. Whereas initially software architecture was primarily associated with the architecture design phase, we now see that the software architecture is treated explicitly during development, product derivation in software product lines, at run-time, and during system evolution. Software architecture as an artifact has been decoupled from a particular lifecycle phase.
With 11 invited submissions from leading researchers and teams of researchers sharing one common characteristic ? all have worked with Dr. Judith Bishop during her long and continuing career as a leader in computer science education and research ? this book reflects on Dr Bishop's outstandingcontribution to computer science. Havingworked at three different universities she now holds a leadership position in the research division of a major software company. The topics covered reflect some of the transitions in her career. The dominant theme is programming languages, with chapters on object oriented programming, real-time programming, component programming and design patterns. Another major and related topic is compilers, with contributions on dataflow analysis, tree rewriting and keyword recognition. Finally, there are some additional chapters on other varied but highly interesting topics including smart homes, mobile systems and teaching computer science."
This book describes a cross-domain architecture and design tools for networked complex systems where application subsystems of different criticality coexist and interact on networked multi-core chips. The architecture leverages multi-core platforms for a hierarchical system perspective of mixed-criticality applications. This system perspective is realized by virtualization to establish security, safety and real-time performance. The impact further includes a reduction of time-to-market, decreased development, deployment and maintenance cost, and the exploitation of the economies of scale through cross-domain components and tools. Describes an end-to-end architecture for hypervisor-level, chip-level, and cluster level. Offers a solution for different types of resources including processors, on-chip communication, off-chip communication, and I/O. Provides a cross-domain approach with examples for wind-power, health-care, and avionics. Introduces hierarchical adaptation strategies for mixed-criticality systems Provides modular verification and certification methods for the seamless integration of mixed-criticality systems. Covers platform technologies, along with a methodology for the development process. Presents an experimental evaluation of technological results in cooperation with industrial partners. The information in this book will be extremely useful to industry leaders who design and manufacture products with distributed embedded systems in mixed-criticality use-cases. It will also benefit suppliers of embedded components or development tools used in this area. As an educational tool, this material can be used to teach students and working professionals in areas including embedded systems, computer networks, system architecture, dependability, real-time systems, and avionics, wind-power and health-care systems.
Knowledge-Based Software Engineering brings together in one place important contributions and up-to-date research results in this important area. Knowledge-Based Software Engineering serves as an excellent reference, providing insight into some of the most important research issues in the field.
The 9th ACIS International Conference on Software Engineering, Artificial Intelligence, Networking, and Parallel/Distributed Computing, held in Phuket Thailand on August 6 - 8, 2008 is aimed at bringing together researchers and scientist, businessmen and entrepreneurs, teachers and students to discuss the numerous fields of computer science, and to share ideas and information in a meaningful way. This publication captures 20 of the conference's most promising papers, and we impatiently await the important contributions that we know these authors will bring to the field.
This book contains a selection of papers presented at the conference on High Performance Software for Nonlinear Optimization (HPSN097) which was held in Ischia, Italy, in June 1997. The rapid progress of computer technologies, including new parallel architec tures, has stimulated a large amount of research devoted to building software environments and defining algorithms able to fully exploit this new computa tional power. In some sense, numerical analysis has to conform itself to the new tools. The impact of parallel computing in nonlinear optimization, which had a slow start at the beginning, seems now to increase at a fast rate, and it is reasonable to expect an even greater acceleration in the future. As with the first HPSNO conference, the goal of the HPSN097 conference was to supply a broad overview of the more recent developments and trends in nonlinear optimization, emphasizing the algorithmic and high performance software aspects. Bringing together new computational methodologies with theoretical ad vances and new computer technologies is an exciting challenge that involves all scientists willing to develop high performance numerical software. This book contains several important contributions from different and com plementary standpoints. Obviously, the articles in the book do not cover all the areas of the conference topic or all the most recent developments, because of the large number of new theoretical and computational ideas of the last few years."
GPS Tracking with Java EE Components: Challenges of Connected Cars highlights how the self-driving car is actually changing the automotive industry, from programing embedded software to hosting services and data crunching, in real time, with really big data. The book analyzes how the challenges of the Self Driving Car (SDC) exceed the limits of a classical GPS Tracking System (GTS.) It provides a guidebook on setting up a tracking system by customizing its components. It also provides an overview of the prototyping and modeling process, and how the reader can modify this process for his or her own software. Every component is introduced in detail and includes a number of design decisions for development. The book introduces Java EE (JEE) Modules, and shows how they can be combined to a customizable GTS, and used as seed components to enrich existing systems with live tracking. The book also explores how to merge tracking and mapping to guide SDCs, and focuses on client server programming to provide useful information. It also discusses the challenges involved with the live coordination of moving cars. This book is designed to aid GTS developers and engineers in the automotive industry. It can also help Java Developers, not only interested in GPS Tracking, but in modern software design from many individual modules. Source code and sample applications will be available on the book's website.
This edited book invites the reader to explore how the latest technologies developed in computational intelligence can be extended and applied to software engineering. Leading experts demonstrate how this recent confluence of software engineering and computational intelligence provides a powerful tool to address the increasing demand for complex applications in diversified areas, the ever-increasing complexity and size of software systems, and the inherently imperfect nature of the information. The presented treatments to software modeling and formal analysis permit the extension of computational intelligence to various phases in software life cycles, such as managing fuzziness resident in the requirements, coping with fuzzy objects and imprecise knowledge, and handling uncertainty encountered in quality prediction.
Formal Equivalence Checking and Design Debugging covers two major topics in design verification: logic equivalence checking and design debugging. The first part of the book reviews the design problems that require logic equivalence checking and describes the underlying technologies that are used to solve them. Some novel approaches to the problems of verifying design revisions after intensive sequential transformations such as retiming are described in detail. The second part of the book gives a thorough survey of previous and recent literature on design error diagnosis and design error correction. This part also provides an in-depth analysis of the algorithms used in two logic debugging software programs, ErrorTracer and AutoFix, developed by the authors. From the Foreword: With the adoption of the static sign-off approach to verifying circuit implementations the application-specific integrated circuit (ASIC) industry will experience the first radical methodological revolution since the adoption of logic synthesis. Equivalence checking is one of the two critical elements of this methodological revolution. This book is timely for either the designer seeking to better understand the mechanics of equivalence checking or for the CAD researcher who wishes to investigate well-motivated research problems such as equivalence checking of retimed designs or error diagnosis in sequential circuits.' Kurt Keutzer, University of California, Berkeley |
You may like...
Artificial Intelligence Technologies and…
Tomayess Issa, Pedro Isaias
Hardcover
R5,697
Discovery Miles 56 970
Handbook of Research on Web Log Analysis
Bernard J Jansen, Amanda Spink, …
Hardcover
R6,726
Discovery Miles 67 260
Evaluating Websites and Web Services
Denis Yannacopoulos, Panagiotis Manolitzas, …
Hardcover
R5,413
Discovery Miles 54 130
|