![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming > Programming languages > General
It is recognized that formal design and verification methods are an important requirement for the attainment of high quality system designs. The field has evolved enormously during the last few years, resulting in the fact that formal design and verification methods are nowadays supported by several tools, both commercial and academic. If different tools and users are to generate and read the same language then it is necessary that the same semantics is assigned by them to all constructs and elements of the language. The current IEEE standard VHDL language reference manual (LRM) tries to define VHDL as well as possible in a descriptive way, explaining the semantics in English. But rigor and clarity are very hard to maintain in a semantics defined in this way, and that has already given rise to many misconceptions and contradictory interpretations. Formal Semantics for VHDL is the first book that puts forward a cohesive set of semantics for the VHDL language. The chapters describe several semantics each based on a different underlying formalism: two of them use Petri nets as target language, and two of them higher order logic. Two use functional concepts, and finally another uses the concept of evolving algebras. Formal Semantics for VHDL is essential reading for researchers in formal methods and can be used as a text for an advanced course on the subject.
This book takes a formal approach to teaching software engineering, using not only UML, but also Object Constraint Language (OCL) for specification and analysis of designed models. Employing technical details typically missing from existing textbooks on software engineering, the author shows how precise specifications lead to static verification of software systems. In addition, data management is given the attention that is required in order to produce a successful software project. Uses constraints in all phases of software development Follows recent developments in software technologies Technical coverage of data management issues and software verification Illustrated throughout to present analysis, specification, implementation and verification of multiple applications Includes end-of-chapter exercises and Instructor Presentation Slides
Welcome to the 5th International Conference on Open Source Systems! It is quite an achievement to reach the five-year mark - that's the sign of a successful enterprise. This annual conference is now being recognized as the primary event for the open source research community, attracting not only high-quality papers, but also building a community around a technical program, a collection of workshops, and (starting this year) a Doctoral Consortium. Reaching this milestone reflects the efforts of many people, including the conference founders, as well as the organizers and participants in the previous conferences. My task has been easy, and has been greatly aided by the hard work of Kevin Crowston and Cornelia Boldyreff, the Program Committee, as well as the Organizing Team led by Bjoern Lundell. All of us are also grateful to our attendees, especially in the difficult economic climate of 2009. We hope the participants found the conference valuable both for its technical content and for its personal networking opportunities. To me, it is interesting to look back over the past five years, not just at this conference, but at the development and acceptance of open source software. Since 2004, the business and commercial side of open source has grown enormously. At that time, there were only a handful of open source businesses, led by RedHat and its Linux distribution. Companies such as MySQL and JBoss were still quite small.
This book is a revised edition of the monograph which appeared under the same title in the series Research Notes in Theoretical Computer Science, Pit man, in 1986. In addition to a general effort to improve typography, English, and presentation, the main novelty of this second edition is the integration of some new material. Part of it is mine (mostly jointly with coauthors). Here is brief guide to these additions. I have augmented the account of categorical combinatory logic with a description of the confluence properties of rewriting systems of categor ical combinators (Hardin, Yokouchi), and of the newly developed cal culi of explicit substitutions (Abadi, Cardelli, Curien, Hardin, Levy, and Rios), which are similar in spirit to the categorical combinatory logic, but are closer to the syntax of A-calculus (Section 1.2). The study of the full abstraction problem for PCF and extensions of it has been enriched with a new full abstraction result: the model of sequential algorithms is fully abstract with respect to an extension of PCF with a control operator (Cartwright, Felleisen, Curien). An order extensional model of error-sensitive sequential algorithms is also fully abstract for a corresponding extension of PCF with a control operator and errors (Sections 2.6 and 4.1). I suggest that sequential algorithms lend themselves to a decomposition of the function spaces that leads to models of linear logic (Lamarche, Curien), and that connects sequentiality with games (Joyal, Blass, Abramsky) (Sections 2.1 and 2.6)."
The Common Intermediate Language (CIL) is "the" core language of .NET. Although .NET developers often use a high-level language (such as C# or VB .NET) to develop their systems, they can use CIL to do "anything" allowed by.NET specificationswhich is not the case for C# and VB .NET. Understanding how CIL works will provide you with a deep, language-independent insight into the core parts of .NET. This knowledge is essential for creating dynamic types, a powerful part of the .NET Framework. In "CIL Programming: Under the Hood of .NET," Jason Bock offers an in-depth tutorial on programming in CIL. First, Bock discusses the basics of .NET assemblies and manifests. He then shows how to create assemblies in .NETincluding the ilasm directives and CIL opcodes, and how these are used to define assemblies, classes, field, methods, and method definitions. Bock also covers the ways in which C#, VB .NET, and other non-Microsoft languages emit CIL, and how they differ. Finally, he reveals how developers can create dynamic assemblies at runtime via the Emitter classes. After reading this guide, you will gain a better understanding of CIL and how to program directly into it. "CIL Programming: Under the Hood of .NET" is a must-have on every .NET developer's desk Table of Contents Language Interoperability ilasm Directives and Attributes CIL Opcodes ilasm and CIL in Practice Debugging CIL .NET Languages and CIL Emitting Types Dynamic Proxies in .NET CIL Tips
Universally acclaimed as the book on garbage collection. A complete and up-to-date revision of the 2012 Garbage Collection Handbook. Thorough coverage of parallel, concurrent and real-time garbage collection algortithms including C4, Garbage First, LXR, Shenandoah, Transactional Sapphire and ZGC, and garbage collection on the GPU. Clear explanation of the trickier aspects of garbage collection, including the interface to the run-time system, handling of finalisation and weak references, and support for dynamic languages. New chapters on energy aware garbage collection, and persistence and garbage collection. The e-book includes more than 40,000 hyperlinks to algorithms, figures, glossary entries, indexed items, original research papers and much more. Backed by a comprehensive online database of over 3,400 garbage collection-related publications
Write Powerful, Modern C++ Code for Scientific, Engineering, and Embedded Applications Discovering Modern C++, Second Edition, will help you master valuable skills for programming with C++ at nearly every level, from "close to the hardware" to high-level abstractions. Updated for C++17 and C++ 20, this intensive introduction teaches C++ using realistic examples from diverse technical problem domains. Drawing on extensive experience teaching C++ to physicists, mathematicians, engineers, and students, Peter Gottschling guides you smoothly to sophisticated approaches based on advanced features. Whatever your programming experience, you'll rapidly master increasingly powerful features, from lambdas to expression and variadic templates. Gottschling also shows you how to apply C++'s libraries: both the Standard Template Library (STL) and scientific libraries for arithmetic, linear algebra, differential equations, and graphs. Step by step, you'll learn to write clear and expressive code using object orientation, generics, metaprogramming, and procedural techniques, and master all the abstractions you need to write high-quality, well-performing software. Quickly master core features: variables, operators, expressions, statements, functions, error handling, I/O, arrays, pointers, references, and more Make the most of classes and object-oriented programming, from constructors/destructors to operator overloading and multiple inheritance Apply advanced generic programming and template-based techniques Use C++'s libraries to write more robust and powerful code more quickly Explore metaprogramming in depth, and master cutting-edge optimization techniques Walk through representative scientific projects, and create your own Register your book for convenient access to downloads, updates, and/or corrections as they become available. See inside book for details.
'One of the best software design books of all time' - BookAuthority Cory Althoff is a self-taught programmer. After a year of self-study, he learned to program well enough to land a job as a software engineer II at eBay. But once he got there, he realised he was severely under-prepared. He was overwhelmed by the amount of things he needed to know but hadn't learned. His journey learning to program, and his experience in first software engineering job were the inspiration for this book. This book is not just about learning to program, although you will learn to code. If you want to program professionally, it is not enough to learn to code; that is why, in addition to helping you learn to program, Althoff also cover the rest of the things you need to know to program professionally that classes and books don't teach you. The Self-taught Programmer is a roadmap, a guide to take you from writing your first Python program to passing your first technical interview. The book is divided into five sections: 1. Learn to program in Python 3 and build your first program. 2. Learn object-oriented programming and create a powerful Python program to get you hooked. 3. Learn to use tools like Git, Bash and regular expressions. Then use your new coding skills to build a web scraper. 4. Study computer science fundamentals like data structures and algorithms. 5. Finish with best coding practices, tips for working with a team and advice on landing a programming job. You can learn to program professionally. The path is there. Will you take it? From the author I spent one year writing The Self-Taught Programmer. It was an exciting and rewarding experience. I treated my book like a software project. After I finished writing it, I created a program to pick out all of the code examples from the book and execute them in Python to make sure all 300+ examples worked properly. Then I wrote software to add line numbers and color to every code example. Finally, I had a group of 200 new programmers 'beta read' the book to identify poorly explained concepts and look for any errors my program missed. I hope you learn as much reading my book as I did writing it. Best of luck with your programming!
"Pro PHP XML and Web Services" is the authoritative guide to using the XML features of PHP 5 and PHP 6. No other book covers XML and Web Services in PHP as deeply as this title. The first four chapters introduce the core concepts of XML required for proficiency, and will bring you up to speed on the terminology and key concepts you need to proceed with the rest of the book. Next, the book explores utilizing XML and Web Services with PHP5. Topics include DOM, SimpleXML, SAX, xmlReader, XSLT, RDF, RSS, WDDX, XML-RPC, REST, SOAP, and UDDI. Author Robert Richards, a major contributor to the PHP XML codebase, is a leading expert in the PHP community. In this book, Richards covers all topics in depth, blending theory with practical examples. You'll find case studies for the most popular web services like Amazon, Google, eBay, and Yahoo. The book also covers XML capabilities, demonstrated through informative examples, in the PEAR libraries.
With "Beginning C: From Novice to Professional, Fourth Edition," you'll come to understand the fundamentals of the C language and learn how to program. All you need is this book and any one of the widely available free or commercial C or C++ compilers, and you'll soon be writing real C programs. You'll learn C from the first principles, using step-by-step working examples that you'll create and execute yourself. This book will increase your programming expertise by guiding you through the development of fully working C applications that use what you've learned in a practical context. You'll also be able to strike out on your own by trying the exercises included at the end of each chapter. Pick up a copy of this book by renowned author, Ivor Horton, because: It is the only beginning-level book to cover the latest ANSI standard in C Is approachable and aimed squarely at people new to C Emphasizes writing code after the first chapter Includes substantial examples relevant to intermediate users
The general markup language XML has played an outstanding role in the mul- ple ways of processing electronic documents, XML being used either in the design of interface structures or as a formal framework for the representation of structure or content-related properties of documents. This book in its 13 chapters discusses aspects of XML-based linguistic information modeling combining: methodological issues, especially with respect to text-related information modeling, applicati- oriented research and issues of formal foundations. The contributions in this book are based on current research in Text Technology, Computational Linguistics and in the international domain of evolving standards for language resources. Rec- rent themes in this book are markup languages, explored from different points of view, and topics of text-related information modeling. These topics have been core areas of the research Unit "Text-technological Information Modeling" (www. te- technology. de) funded from 2002 to 2009 by the German Research Foundation (DFG). Positions developed in this book could also bene t from the presentations and discussion at the conference "Modelling Linguistic Information Resources" at the Center for Interdisciplinary Research (Zentrum fur .. interdisziplinare .. Forschung, ZiF) at Bielefeld, a center for advanced studies known for its international and interdisciplinary meetings and research. The editors would like to thank the DFG and ZiF for their nancial support, the publisher, the series editors, the reviewers and those people that helped to prepare the manuscript, especially Carolin Kram, Nils Diewald, Jens Stegmann and Peter M. Fischer and last but not least, all of the authors.
An excerpt from "The Unwanted Gift" Stuart Hart awoke to find that he was floating alongside the couch. His eyes were wide open. Their view --- the ceiling. The bright bulb caused his pupils to constrict - everything was delicately sharp. He was lying on his back - torso - legs - arms - head - all in one plane, as if someone or something had pulled the couch out from under him, but the couch had not moved. It was still in its place against the wall. Stuart Hart had moved He was suspended in space. During his sleep, the sentinel of consciousness had failed to detect some force that changed him. Something had insinuated itself into his body. The sensation of self or solidity was gone. Did he still exist? Had he died in sleep? Was he now in transition from body to spirit? Deliberately, he moved one hand towards his face, fearfully expecting not to see that hand, nor to feel his face when he touched it. He "was" still there - a body with physical properties.
In a quiet and cozy corner of the cosmos, God prepared a home for his children. On this tiny world, he built a paradise for them in which to be nurtured and raised. With his angel's mentoring and encouragement, his children and all the world would enjoy Heaven on Earth. But, there was one angel who did not like that plan. The servant was jealous of his master and wished not only to capture his kingdom, but his queen, as well. Coveting her so and being unable to have her. Thinking of nothing except to possess the throne of her lord and all the privileges of it. Unfilled desire and jealous frustration. These things drove the angel to anger. And his anger drove him mad.
For real-time systems, the worst-case execution time (WCET) is the key objective to be considered. Traditionally, code for real-time systems is generated without taking this objective into account and the WCET is computed only after code generation. Worst-Case Execution Time Aware Compilation Techniques for Real-Time Systems presents the first comprehensive approach integrating WCET considerations into the code generation process. Based on the proposed reconciliation between a compiler and a timing analyzer, a wide range of novel optimization techniques is provided. Among others, the techniques cover source code and assembly level optimizations, exploit machine learning techniques and address the design of modern systems that have to meet multiple objectives. Using these optimizations, the WCET of real-time applications can be reduced by about 30% to 45% on the average. This opens opportunities for decreasing clock speeds, costs and energy consumption of embedded processors. The proposed techniques can be used for all types real-time systems, including automotive and avionics IT systems.
This book represents an attempt to treat three aspects of digital systems, design, prototyping and customization, in an integrated manner using two major technologies: VHSIC Hardware Description Language (VHDL) as a modeling and specification tool, and Field-Programmable Logic Devices (FPLDs) as an implementation technology. They together make a very powerful combination for complex digital systems rapid design and prototyping as the important steps towards manufacturing, or, in the case of feasible quantities, they also provide fast system manufacturing. Combining these two technologies makes possible implementation of very complex digital systems at the desk. VHDL has become a standard tool to capture features of digital systems in a form of behavioral, dataflow or structural models providing a high degree of flexibility. When augmented by a good simulator, VHDL enables extensive verification of features of the system under design, reducing uncertainties at the latter phases of design process. As such, it becomes an unavoidable modeling tool to model digital systems at various levels of abstraction.
An Introduction to R and Python for Data Analysis helps teach students to code in both R and Python simultaneously. As both R and Python can be used in similar manners, it is useful and efficient to learn both at the same time, helping lecturers and students to teach and learn more, save time, whilst reinforcing the shared concepts and differences of the systems. This tandem learning is highly useful for students, helping them to become literate in both languages, and develop skills which will be handy after their studies. This book presumes no prior experience with computing, and is intended to be used by students from a variety of backgrounds. The side-by-side formatting of this book helps introductory graduate students quickly grasp the basics of R and Python, with the exercises providing helping them to teach themselves the skills they will need upon the completion of their course, as employers now ask for competency in both R and Python. Teachers and lecturers will also find this book useful in their teaching, providing a singular work to help ensure their students are well trained in both computer languages. All data for exercises can be found here: https://github.com/tbrown122387/r_and_python_book/tree/master/data. Key features: - Teaches R and Python in a "side-by-side" way. - Examples are tailored to aspiring data scientists and statisticians, not software engineers. - Designed for introductory graduate students. - Does not assume any mathematical background.
Many times, web services standards do not explicitly address core issues specific to the financial industrywhich makes it difficult to implement standards-compliant systems. But "Web Services in Finance "will bridge the gap in standards awareness. And you will acquire the skills to develop secure applications quickly. If you are a .NET or J2EE developer working in the financial industry, currently migrating applications to become Web services, or writing new Web services, then this book is your ideal companion! The authors thoroughly discuss crucial topics like data representation, messaging, security, privacy, management, monitoring, and more. What's more: the provided examples and API reviews will help you swiftly reach your goals. Table of Contents Introduction to Web Services Enterprise Systems Data Representation Messaging Description and Data Format Discovery and Advertising Alternative Transports Security Quality of Service Conversations, Workflows, and Transactions
The proceedings represent the state of knowledge in the area of algorithmic differentiation (AD). The 31 contributed papers presented at the AD2012 conference cover the application of AD to many areas in science and engineering as well as aspects of AD theory and its implementation in tools. For all papers the referees, selected from the program committee and the greater community, as well as the editors have emphasized accessibility of the presented ideas also to non-AD experts. In the AD tools arena new implementations are introduced covering, for example, Java and graphical modeling environments or join the set of existing tools for Fortran. New developments in AD algorithms target the efficiency of matrix-operation derivatives, detection and exploitation of sparsity, partial separability, the treatment of nonsmooth functions, and other high-level mathematical aspects of the numerical computations to be differentiated. Applications stem from the Earth sciences, nuclear engineering, fluid dynamics, and chemistry, to name just a few. In many cases the applications in a given area of science or engineering share characteristics that require specific approaches to enable AD capabilities or provide an opportunity for efficiency gains in the derivative computation. The description of these characteristics and of the techniques for successfully using AD should make the proceedings a valuable source of information for users of AD tools.
Renowned expert of the occult, Dr. Walter Prine, is summoned to a town in West Virginia to interview nine survivors of an unimaginable holocaust. They claimed that an angel had killed all but nine for failure to deliver a man named Jericho Black. As the sessions intensify, Dr. Prine learns that he is the true target of the vengeful angel, for hiding one of the most precious possessions in existence: An apple from the Garden of Eden taken from a world without sin. But Prine suffers from amnesia at the hand of God as punishment for his treachery and does not remember anything about the apple. Prine soon finds himself in the middle of a search between Heaven and Hell to find the omnipotent fruit that might end the war between good and evil forever. But the cost may be the end of life itself.
This volume gives an overview of the state-of-the-art in system-level design trade-off explorations for concurrent tasks running on embedded heterogeneous multiple processors. The targeted application domain covers complex embedded real-time multi-media and communication applications. Many of these applications are concurrent in the sense that multiple subsystems can be running simultaneously. Also, these applications are so dynamic at run-time that the designs based on the worst case execution times are inefficient in terms of resource allocation (e.g., energy budgets). A novel systematical approach is clearly necessary in the area of system-level design for the embedded systems where those concurrent and dynamic applications are mapped. This material is mainly based on research at IMEC and its international university network partners in this area in the period 1997-2006.
After a slow and somewhat tentative beginning, machine vision systems are now finding widespread use in industry. So far, there have been four clearly discernible phases in their development, based upon the types of images processed and how that processing is performed: (1) Binary (two level) images, processing in software (2) Grey-scale images, processing in software (3) Binary or grey-scale images processed in fast, special-purpose hardware (4) Coloured/multi-spectral images Third-generation vision systems are now commonplace, although a large number of binary and software-based grey-scale processing systems are still being sold. At the moment, colour image processing is commercially much less significant than the other three and this situation may well remain for some time, since many industrial artifacts are nearly monochrome and the use of colour increases the cost of the equipment significantly. A great deal of colour image processing is a straightforward extension of standard grey-scale methods. Industrial applications of machine vision systems can also be sub divided, this time into two main areas, which have largely retained distinct identities: (i) Automated Visual Inspection (A VI) (ii) Robot Vision (RV) This book is about a fifth generation of industrial vision systems, in which this distinction, based on applications, is blurred and the processing is marked by being much smarter (i. e. more "intelligent") than in the other four generations."
Case-based reasoning means reasoning based on remembering previous experiences. A reasoner using old experiences (cases) might use those cases to suggest solutions to problems, to point out potential problems with a solution being computed, to interpret a new situation and make predictions about what might happen, or to create arguments justifying some conclusion. A case-based reasoner solves new problems by remembering old situations and adapting their solutions. It interprets new situations by remembering old similar situations and comparing and contrasting the new one to old ones to see where it fits best. Case-based reasoning combines reasoning with learning. It spans the whole reasoning cycle. A situation is experienced. Old situations are used to understand it. Old situations are used to solve a problem (if there is one to be solved). Then the new situation is inserted into memory alongside the cases it used for reasoning, to be used another time. The key to this reasoning method, then, is remembering. Remembering has two parts: integrating cases or experiences into memory when they happen and recalling them in appropriate situations later on. The case-based reasoning community calls this related set of issues the indexing problem. In broad terms, it means finding in memory the experience closest to a new situation. In narrower terms, it can be described as a two-part problem: assigning indexes or labels to experiences when they are put into memory that describe the situations to which they are applicable, so that they can be recalled later; and at recall time, elaborating the new situation in enough detail so that the indexes it would have if it were in the memory are identified. Case-Based Learning is an edited volume of original research comprising invited contributions by leading workers. This work has also been published as a special issues of MACHINE LEARNING, Volume 10, No. 3.
This study centers on issues of marginality and monstrosity in medieval England. In the middle ages, geography was viewed as divinely ordered, so Britain's location at the periphery of the inhabitable world caused anxiety among its inhabitants. Far from the world's holy center, the geographic margins were considered monstrous. Medieval geography, for centuries scorned as crude, is now the subject of several careful studies. Monsters have likewise been the subject of recent attention in the growing field of "monster studies," though few works situate these creatures firmly in their specific historical contexts. This study sits at the crossroads of these two discourses (geography and monstrosity), treated separately in the established scholarship but inseparable in the minds of medieval authors and artists. |
You may like...
Introducing Delphi Programming - Theory…
John Barrow, Linda Miller, …
Paperback
(1)R751 Discovery Miles 7 510
Introduction to Computational Economics…
Hans Fehr, Fabian Kindermann
Hardcover
R4,258
Discovery Miles 42 580
Advanced Visual Basic 6 - Power…
Matthew Curland, Gary Clarke
Paperback
R1,273
Discovery Miles 12 730
Dark Silicon and Future On-chip Systems…
Suyel Namasudra, Hamid Sarbazi-Azad
Hardcover
R3,940
Discovery Miles 39 400
|