![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming > Programming languages
This book provides a superb introduction to and overview of the MIT PI System for custom VLSI placement and routing. Alan Sher man has done an excellent job of collecting and clearly presenting material that was previously available only in various theses, confer ence papers, and memoranda. He has provided here a balanced and comprehensive presentation of the key ideas and techniques used in PI, discussing part of his own Ph. D. work (primarily on the place ment problem) in the context of the overall design of PI and the contributions of the many other PI team members. I began the PI Project in 1981 after learning first-hand how dif ficult it is to manually place modules and route interconnections in a custom VLSI chip. In 1980 Adi Shamir, Leonard Adleman, and I designed a custom VLSI chip for performing RSA encryp tion/decryption [226]. I became fascinated with the combinatorial and algorithmic questions arising in placement and routing, and be gan active research in these areas. The PI Project was started in the belief that many of the most interesting research issues would arise during an actual implementation effort, and secondarily in the hope that a practically useful tool might result. The belief was well-founded, but I had underestimated the difficulty of building a large easily-used software tool for a complex domain; the PI soft ware should be considered as a prototype implementation validating the design choices made.
This book constitutes the refereed proceedings of the 11th International Symposium on Functional and Logic Programming, FLOPS 2012, held in Kobe, Japan, in May 2012. The 19 research papers and 3 system demonstrations presented in this volume were carefully reviewed and selected from 39 submissions. They deal with declarative programming, including functional programming and logic programming.
This book had its genesis in the following piece of computer mail: From allegra!joan-b Tue Dec 18 89:15:54 1984 To: sola!hjb Subj ect: 1 i spm Hank, I've been talking with Mark Plotnik and Bill Gale about asking you to conduct a basic course on using the lisp machine. Mark, for instance, would really like to cover basics like the flavor system, etc. , so he could start doing his own programming without a lot of trial and error, and Bill and I would be interested in this, too. I'm quite sure that Mark Jones, Bruce, Eric and Van would also be really interested. Would you like to do it? Bill has let me know that if you'd care to set something up, he's free to meet with us anytime this week or next (although I'll only be here on Wed. next week) so we can come up with a plan. What do you think? Joan. xiv Lisp Lore (All the people and computers mentioned above work at AT&T Bell Laboratories, in Murray Hill, New Jersey. ) I agreed, with some trepidation, to try teaching such a course. It wasn't clear how I was going to explain the Lisp Machine environment to a few dozen beginners when at the time I felt I was scarcely able to keep myself afloat. Particularly since many of the "beginners" had PhD's in computer science and a decade or two of programming experience.
Logical concepts and methods are of growing importance in many areas of computer science. The proofs-as-programs paradigm and the wide acceptance of Prolog show this clearly. The logical notion of a formal proof in various constructive systems can be viewed as a very explicit way to describe a computation procedure. Also conversely, the development of logical systems has been influenced by accumulating knowledge on rewriting and unification techniques. This volume contains a series of lectures by leading researchers giving a presentation of new ideas on the impact of the concept of a formal proof on computation theory. The subjects covered are: specification and abstract data types, proving techniques, constructive methods, linear logic, and concurrency and logic.
The topic of logic programming and databases. has gained in creasing interest in recent years. Several events have marked the rapid evolution of this field: the selection, by the Japanese Fifth Generation Project, of Prolog and of the relational data model as the basis for the development of new machine archi tectures; the focusing of research in database theory on logic queries and on recursive query processing; and the pragmatic, application-oriented development of expert database systems and of knowledge-base systems. As a result, an enormous amount of work has been produced in the recent literature, coupled with the spontaneous growth of several advanced projects in this area. The goal of this book is to present a systematic overview of a rapidly evolving discipline, which is presently not described with the same approach in other books. We intend to introduce stu dents and researchers to this new discipline; thus we use a plain, tutorial style, and complement the description of algorithms with examples and exercises. We attempt to achieve a balance be tween theoretical foundations and technological issues; thus we present a careful introduction to the new language Datalog, but we also focus on the efficient interfacing of logic programming formalisms (such as Prolog and Datalog) with large databases.
This book constitutes the thoroughly refereed post-proceedings of the 23rd International Workshop on Languages and Compilers for Parallel Computing, LCPC 2010, held in Houston, TX, USA, in October 2010. The 18 revised full papers presented were carefully reviewed and selected from 47 submissions. The scope of the workshop spans foundational results and practical experience, and targets all classes of parallel platforms including concurrent, multithreaded, multicore, accelerated, multiprocessor, and cluster systems
This book constitutes the proceedings of the 15th International Conference on Foundations of Software Science and Computational Structures, FOSSACS 2012, held as part of the joint European Conference on Theory and Practice of Software, ETAPS 2012, which took place in Tallinn, Estonia, in March/April 2012. The 29 papers presented in this book together with two invited talks in full paper length were carefully reviewed and selected from 100 full paper submissions. The papers deal with theories and methods to support analysis, synthesis, transformation and verification of programs and software systems.
Introduction The goal of this book is to introduce XML to a bioinformatics audience. It does so by introducing the fundamentals of XML, Document Type De?nitions (DTDs), XML Namespaces, XML Schema, and XML parsing, and illustrating these concepts with speci?c bioinformatics case studies. The book does not assume any previous knowledge of XML and is geared toward those who want a solid introduction to fundamental XML concepts. The book is divided into nine chapters: Chapter 1: Introduction to XML for Bioinformatics. This chapter provides an introduction to XML and describes the use of XML in biological data exchange. A bird's-eye view of our ?rst case study, the Distributed Annotation System (DAS), is provided and we examine a sample DAS XML document. The chapter concludes with a discussion of the pros and cons of using XML in bioinformatic applications. Chapter 2: Fundamentals of XML and BSML. This chapter introduces the fundamental concepts of XML and the Bioinformatic Sequence Markup Language (BSML). We explore the origins of XML, de?ne basic rules for XML document structure, and introduce XML Na- spaces. We also explore several sample BSML documents and visualize these documents in the TM Rescentris Genomic Workspace Viewer.
"Oil is the problem. Cars are the solution."
Beginning with an informal introduction to language equations, this book presents a framework for a general theory for solving systems of equations and relations between languages. Classical language equations, generalized derivatives, Boolean language equations, and implicit equations are presented systematically. An exploration of mixed systems and open problems rounds out the presentation.
You might well wonder why TFPC in Practice is a part of the Monographs in Visualization series. However, if you really think about typesetting, especially fine typesetting, you soon realize that in large part it is a visual art as well as a science. 'lEX's algorithms produce in almost all cases aesthetic results of the highest quality. On the other hand, occasionally one may want to insert some additional space before a subscript or superscript, or one may want to adjust the vertical spacing in a fraction. Fortunately Donald Knuth, the author of 'lEX, allows one to program such corrections easily where needed. The four volumes of Stephan von Bechtolsheim's long awaited TFPC in Prac tice present a comprehensive view of 'lEX. His thorough discussion of each aspect of 'lEX is liberally laced with cogent illustrative examples. Many of these exam ples represent complete, ready to use macros that enhance the capabilities of 'lEX. These examples are of particular interest to both the typesetter and the 'lEX programmer. The typesetter can often solve an immediate problem by ei ther using one of the examples directly or by making minor changes to adapt it to the problem at hand. The 'lEX programmer can use the examples, along with Stephan's detailed discussion, to increase both the depth and breadth of his or her knowledge of 'lEX. The value of the text is further enhanced by Stephan's concerted effort to explain the reasoning behind each topic or example."
System-on-Chip Methodologies & Design Languages brings together a selection of the best papers from three international electronic design language conferences in 2000. The conferences are the Hardware Description Language Conference and Exhibition (HDLCon), held in the Silicon Valley area of USA; the Forum on Design Languages (FDL), held in Europe; and the Asia Pacific Chip Design Language (APChDL) Conference. The papers cover a range of topics, including design methods, specification and modeling languages, tool issues, formal verification, simulation and synthesis. The results presented in these papers will help researchers and practicing engineers keep abreast of developments in this rapidly evolving field.
Contributions on UML address the application of UML in the
specification of embedded HW/SW systems. C-Based System Design
embraces the modeling of operating systems, modeling with different
models of computation, generation of test patterns, and experiences
from case studies with SystemC. Analog and Mixed-Signal Systems
covers rules for solving general modeling problems in VHDL-AMS,
modeling of multi-nature systems, synthesis, and modeling of
Mixed-Signal Systems with SystemC. Languages for formal methods are
addressed by contributions on formal specification and refinement
of hybrid, embedded and real-time stems.
Duration calculus constitutes a formal approach to the development of real-time systems; as an interval logic with special features for expressing and analyzing time durations of states in real-time systems, it allows for representing and formally reasoning about requirements and designs at an appropriate level of abstraction. This book presents the logical foundations of duration calculus in a coherent and thorough manner. Through selective case studies it explains how duration calculus can be applied to the formal specification and verification of real-time systems. The book also contains an extensive survey of the current research in this field. The material included in this book has been used for graduate and postgraduate courses, while it is also suitable for experienced researchers and professionals.
This book brings together experts to discuss relevant results in software process modeling, and expresses their personal view of this field. It is designed for a professional audience of researchers and practitioners in industry, and graduate-level students.
Reasoning under uncertainty is always based on a specified language or for malism, including its particular syntax and semantics, but also on its associated inference mechanism. In the present volume of the handbook the last aspect, the algorithmic aspects of uncertainty calculi are presented. Theory has suffi ciently advanced to unfold some generally applicable fundamental structures and methods. On the other hand, particular features of specific formalisms and ap proaches to uncertainty of course still influence strongly the computational meth ods to be used. Both general as well as specific methods are included in this volume. Broadly speaking, symbolic or logical approaches to uncertainty and nu merical approaches are often distinguished. Although this distinction is somewhat misleading, it is used as a means to structure the present volume. This is even to some degree reflected in the two first chapters, which treat fundamental, general methods of computation in systems designed to represent uncertainty. It has been noted early by Shenoy and Shafer, that computations in different domains have an underlying common structure. Essentially pieces of knowledge or information are to be combined together and then focused on some particular question or domain. This can be captured in an algebraic structure called valuation algebra which is described in the first chapter. Here the basic operations of combination and focus ing (marginalization) of knowledge and information is modeled abstractly subject to simple axioms."
Electronic Chips & Systems Design Languagesoutlines and describes the latest advances in design languages. The challenge of System on a Chip (SOC) design requires designers to work in a multi-lingual environment which is becoming increasingly difficult to master. It is therefore crucial for them to learn, almost in real time, from the experiences of their colleagues in the use of design languages and how these languages have become more advanced to cope with system design. System designers, as well as students willing to become system designers, often do not have the time to attend all scientific events where they could learn the necessary information. This book will bring them a selected digest of the best contributions and industry strength case studies. All the levels of abstraction that are relevant, from the informal user requirements down to the implementation specifications, are addressed by different contributors. The author, together with colleague authors who provide valuable additional experience, presents examples of actual industrial world applications. Furthermore the academic concepts presented in this book provide excellent theories to student readers and the concepts described are up to date and in so doing provide most suitable root information for Ph.D. postgraduates.
Written by the members of the IFIP Working Group 2.3 (Programming Methodology) this text constitutes an exciting reference on the front-line of research activity in programming methodology. The range of subjects reflects the current interests of the members, and will offer insightful and controversial opinions on modern programming methods and practice. The material is arranged in thematic sections, each one introduced by a problem which epitomizes the spirit of that topic. The exemplary problem will encourage vigorous discussion and will form the basis for an introduction/tutorial for its section.
The text contains a detailed and current presentation of the program analyses and transformations that extract the flow of data in computer memory systems. The emphasis is on a framework for the optimization of code for imperative programs and greater computer systems efficiency. In addition, the author shows that correctness of program transformations is guaranteed by the conservation of data flow. Professionals and researchers in software engineering, computer engineering, program design analysis, and compiler design will benefit from its presentation of data-flow methods and memory optimization of compilers.
Goal Directed Proof Theory presents a uniform and coherent methodology for automated deduction in non-classical logics, the relevance of which to computer science is now widely acknowledged. The methodology is based on goal-directed provability. It is a generalization of the logic programming style of deduction, and it is particularly favourable for proof search. The methodology is applied for the first time in a uniform way to a wide range of non-classical systems, covering intuitionistic, intermediate, modal and substructural logics. The book can also be used as an introduction to these logical systems form a procedural perspective. Readership: Computer scientists, mathematicians and philosophers, and anyone interested in the automation of reasoning based on non-classical logics. The book is suitable for self study, its only prerequisite being some elementary knowledge of logic and proof theory.
As information technologies become increasingly distributed and accessible to larger number of people and as commercial and government organizations are challenged to scale their applications and services to larger market shares, while reducing costs, there is demand for software methodologies and appli- tions to provide the following features: Richer application end-to-end functionality; Reduction of human involvement in the design and deployment of the software; Flexibility of software behaviour; and Reuse and composition of existing software applications and systems in novel or adaptive ways. When designing new distributed software systems, the above broad requi- ments and their translation into implementations are typically addressed by partial complementarities and overlapping technologies and this situation gives rise to significant software engineering challenges. Some of the challenges that may arise are: determining the components that the distributed applications should contain, organizing the application components, and determining the assumptions that one needs to make in order to implement distributed scalable and flexible applications, etc.
This work is Volume II of a two-volume monograph on the theory of deterministic parsing of context-free grammars. Volume I, "Languages and Parsing" (Chapters 1 to 5), was an introduction to the basic concepts of formal language theory and context-free parsing. Volume II (Chapters 6 to 10) contains a thorough treat ment of the theory of the two most important deterministic parsing methods: LR(k) and LL(k) parsing. Volume II is a continuation of Volume I; together these two volumes form an integrated work, with chapters, theorems, lemmas, etc. numbered consecutively. Volume II begins with Chapter 6 in which the classical con structions pertaining to LR(k) parsing are presented. These include the canonical LR(k) parser, and its reduced variants such as the LALR(k) parser and the SLR(k) parser. The grammarclasses for which these parsers are deterministic are called LR(k) grammars, LALR(k) grammars and SLR(k) grammars; properties of these grammars are also investigated in Chapter 6. A great deal of attention is paid to the rigorous development of the theory: detailed mathematical proofs are provided for most of the results presented."
Computer-Aided Reasoning: ACL2 Case Studies illustrates how the
computer-aided reasoning system ACL2 can be used in productive and
innovative ways to design, build, and maintain hardware and
software systems. Included here are technical papers written by
twenty-one contributors that report on self-contained case studies,
some of which are sanitized industrial projects. The papers deal
with a wide variety of ideas, including floating-point arithmetic,
microprocessor simulation, model checking, symbolic trajectory
evaluation, compilation, proof checking, real analysis, and several
others.
This book constitutes the thoroughly refereed post-proceedings of
the 4th International Conference on Software Language Engineering,
SLE 2011, held in Braga, Portugal, in July 2011.
This Festschrift volume is published in honor of Dexter Kozen on the occasion of his 60th birthday. Dexter Kozen has been a leader in the development of Kleene Algebras (KAs). The contributions in this volume reflect the breadth of his work and influence. The volume includes 19 full papers related to Dexter Kozen's research. They deal with coalgebraic methods, congruence closure; the completeness of various programming logics; decision procedure for logics; alternation; algorithms and complexity; and programming languages and program analysis. The second part of this volume includes laudatios from several collaborators, students and friends, including the members of his current band. |
You may like...
Perspectives on Element Theory
Sabrina Bendjaballah, Ali Tifrit, …
Hardcover
R3,458
Discovery Miles 34 580
Speech Production and Perception
Susanne Fuchs, Amelie 0000-0002-2098 Rochet-Capellan, …
Hardcover
R1,193
Discovery Miles 11 930
Online Learning and Assessment in Higher…
Robyn Benson, Charlotte Brack
Paperback
R1,554
Discovery Miles 15 540
Letters are Characters - A Play-Based…
Caroline Wilcox Ugurlu
Hardcover
A Process Algebraic Approach to Software…
Alessandro Aldini, Marco Bernardo, …
Hardcover
R1,577
Discovery Miles 15 770
|