![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming > Programming languages
Models of Sharing Graphs presents a sound mathematical basis for reasoning about models of computation involving shared resources, including graph rewriting systems, denotational semantics and concurrency theory. An algebraic approach, based on the language of category theory, is taken throughout this work, which enables the author to describe several aspects of the notion of sharing in a systematic way. In particular, a novel account of recursive computation created from cyclic sharing is developed using this framework.
This book constitutes the thoroughly refereed post-workshop proceedings of the 8th International Workshop on Declarative Agent Languages and Technologies, DALT 2010, held in Toronto, Canada, on May 10, 2010, as a satellite workshop of the 9th International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS 2010. The 7 revised full papers presented together with 4 invited lectures were carefully selected during two rounds of reviewing and improvement from 24 initial submissions. DALT aims to make formal methods and declarative technologies and approaches available to and understood by a broader segment of the multi-agent research community; the papers are organized in topical sections on BDI rational agents, communication, coordination and negotiation, as well as social aspects and control systems.
Rule-Based Programming is a broad presentation of the rule-based programming method with many example programs showing the strengths of the rule-based approach. The rule-based approach has been used extensively in the development of artificial intelligence systems, such as expert systems and machine learning. This rule-based programming technique has been applied in such diverse fields as medical diagnostic systems, insurance and banking systems, as well as automated design and configuration systems. Rule-based programming is also helpful in bridging the semantic gap between an application and a program, allowing domain specialists to understand programs and participate more closely in their development. Over sixty programs are presented and all programs are available from an ftp site. Many of these programs are presented in several versions allowing the reader to see how realistic programs are elaborated from `back of envelope' models. Metaprogramming is also presented as a technique for bridging the `semantic gap'. Rule-Based Programming will be of interest to programmers, systems analysts and other developers of expert systems as well as to researchers and practitioners in artificial intelligence, computer science professionals and educators.
This book constitutes the refereed proceedings of the 5th International Workshop on Reachability Problems, RP 2011, held in Genoa, Italy, in September 2011. The 16 papers presented together with 4 invited talks were carefully reviewed and selected from 24 submissions. The workshop deals with reachability problems that appear in algebraic structures, computational models, hybrid systems, logic, and verification. Reachability is a fundamental problem that appears in several different contexts: finite- and infinite-state concurrent systems, computational models like cellular automata and Petri nets, decision procedures for classical, modal and temporal logic, program analysis, discrete and continuous systems, time critical systems, and open systems modelled as games.
by Luea Cardelli Ever since Strachey's work in the 1960's, polymorphism has been classified into the parametric and overloading varieties. Parametric polymorphism has been the subject of extensive study for over two decades. Overloading, on the other hand, has often been considered too ad hoc to deserve much attention even though it has been, in some form, an ingredient of virtually every programming lan guage (much more so than parametric polymorphism). With the introduction of object-oriented languages, and in particular with multiple-dispatch object-oriented languages, overloading has become less of a programming convenience and more of a fundamental feature in need of proper explanation. This book provides a compelling framework for the study of run-time over loading and of its interactions with subtyping and with parametric polymorphism. The book also describes applications to object-oriented programming. This new framework is motivated by the relatively recent spread of programming languages that are entirely based on run-time overloading; this fact probably explains why this subject was not investigated earlier. Once properly understood, overloading reveals itself relevant also to the study of older and more conventional (single dispatch) object-oriented languages, clarifying delicate issues of covariance and contravariance of method types, and of run-time type analysis. In the final chapters, a synthesis is made between parametric and overloading polymorphism."
TEX has always been regarded as the most elegant and powerful system for computer typesetting. However, its widespread use, beyond academia, was hampered by its complexity. Recently, fairly good TEX implementations have come out for PCs putting TEX on the desks of many people: writers, designers, desktop publishers, engineers, and consequently, the interest in TEX has surged. What is needed at this point is a book that teaches step-by-step how to use TEX, illustrating each step by meaningful examples. This is exactly what S.v. Bechtolsheim's book does. It is a tutorial and guide for the first-time users of TEX, as well as a reference for the most experienced "TEXpert." TEX in Practice will appear as a four volume set, starting with volume 1 "Basics," followed by volume 2 "Paragraphs, Math and " "Fonts," volume 3 "Tokens, Macros" and volume 4 "Output Routines, " "Tables." TEX in Practice will be an indispensable reference for the TEX community and a guide through the first steps for the TEX novice.
This Festschrift volume, published in honor of Carolyn Talcott on the occasion of her 70th birthday, contains a collection of papers presented at a symposium held in Menlo Park, California, USA, in November 2011. Carolyn Talcott is a leading researcher and mentor of international renown among computer scientists. She has made key contributions to a number of areas of computer science including: semantics and verification of progamming languages; foundations of actor-based systems; middleware, meta-architectures, and systems; Maude and rewriting logic; and computational biology. The 21 papers presented are organized in topical sections named: Essays on Carolyn Talcott; actors and programming languages; cyberphysical systems; middleware and meta-architectures; formal methods and reasoning tools; and computational biology.
This book constitutes the refereed proceedings of the 9th International Conference on Formal Modeling and Analysis of Timed Systems, FORMATS 2011, held in Aalborg, Denmark, in September 2011. The 20 revised full papers presented together with three invited talks were carefully reviewed and selected from 43 submissions. The papers are organized in topical sections on probabilistic methods, robustness, games, verification and testing, verification, hybrid systems, and applications.
Not long ago" Dennis Merritt wrote one of the best books that I know of about implementing expert systems in Prolog, and I was very glad he published it in our series. The only problem is there are still some unfortunate people around who do not know Prolog and are not sufficiently prepared either to read Merritt's book, or to use this extremely productive language, be it for knowledge-based work or even for everyday programming. Possibly this last statement may surprise you if you were under the impression that Prolog was an "artificial intelligence language" with very limited application potential. Please believe this editor's statement that quite the opposite is true: for at least four years, I have been using Prolog for every programming task in which I am given the option of choosing the language. Therefore, I 'am indeed happy that Dennis Merritt has written another good book on my language of choice, and that it meets the high standard he set with his prior book, Building Expert Systems in Prolog. All that remains for me to do is to wish you success and enjoyment when taking off on your Adventure in Prolog.
Now available in paperback With Beginning C: From Novice to Professional, Fourth Edition, you'll come to understand the fundamentals of the C language and learn how to program. All you need is this book and any one of the widely available free or commercial C or C++ compilers, and you'll soon be writing real C programs. You'll learn C from the first principles, using step-by-step working examples that you'll create and execute yourself. This book will increase your programming expertise by guiding you through the development of fully working C applications that use what you've learned in a practical context. You'll also be able to strike out on your own by trying the exercises included at the end of each chapter. Pick up a copy of this book by renowned author, Ivor Horton, because: It is the only beginning-level book to cover the latest ANSI standard in C Is approachable and aimed squarely at people new to C Emphasizes writing code after the first chapter Includes substantial examples relevant to intermediate users
Intended for first- or second-year undergraduates, this introduction to discrete mathematics covers the usual topics of such a course, but applies constructivist principles that promote - indeed, require - active participation by the student. Working with the programming language ISETL, whose syntax is close to that of standard mathematical language, the student constructs the concepts in her or his mind as a result of constructing them on the computer in the syntax of ISETL. This dramatically different approach allows students to attempt to discover concepts in a "Socratic" dialog with the computer. The discussion avoids the formal "definition-theorem" approach and promotes active involvement by the reader by its questioning style. An instructor using this text can expect a lively class whose students develop a deep conceptual understanding rather than simply manipulative skills. Topics covered in this book include: the propositional calculus, operations on sets, basic counting methods, predicate calculus, relations, graphs, functions, and mathematical induction.
R is the amazing, free, open-access software package for scientific graphs and calculations used by scientists worldwide. The R Student Companion is a student-oriented manual describing how to use R in high school and college science and mathematics courses. Written for beginners in scientific computation, the book assumes the reader has just some high school algebra and has no computer programming background. The author presents applications drawn from all sciences and social sciences and includes the most often used features of R in an appendix. In addition, each chapter provides a set of computational challenges: exercises in R calculations that are designed to be performed alone or in groups. Several of the chapters explore algebra concepts that are highly useful in scientific applications, such as quadratic equations, systems of linear equations, trigonometric functions, and exponential functions. Each chapter provides an instructional review of the algebra concept, followed by a hands-on guide to performing calculations and graphing in R. R is intuitive, even fun. Fantastic, publication-quality graphs of data, equations, or both can be produced with little effort. By integrating mathematical computation and scientific illustration early in a student s development, R use can enhance one's understanding of even the most difficult scientific concepts. While R has gained a strong reputation as a package for statistical analysis, The R Student Companion approaches R more completely as a comprehensive tool for scientific computing and graphing.
These proceedings containall the papers that werepresented at the 4th Inter- tional Conference on Language and Automata Theory and Applications (LATA 2010), held in Trier, Germany, during May 24-28, 2010. The scope of LATA is rather broad, including: algebraic language theory; algorithms on automata and words; automata and logic; automata for system analysis and program veri?cation; automata, concurrency and Petri nets; cel- lar automata; combinatorics on words; computability; computational compl- ity; computer linguistics; data and image compression; decidability questions on words and languages; descriptional complexity; DNA and other models of bio-inspired computing; document engineering; foundations of ?nite state te- nology; fuzzy and rough languages; grammars (Chomsky hierarchy, contextual, multidimensional, uni?cation, categorial, etc. ); grammars and automata arc- tectures; grammatical inference and algorithmic learning; graphs and graph transformation; language varieties and semigroups; language-based cryptog- phy; language-theoretic foundations of arti?cial intelligence and arti?cial life; neuralnetworks;parallelandregulatedrewriting;parsing;patternmatching and pattern recognition; patterns and codes; power series; quantum, chemical and optical computing; semantics; string and combinatorial issues in computational biology and bioinformatics; symbolic dynamics; term rewriting; text algorithms; textretrieval;transducers;trees, treelanguagesandtreemachines;andweighted machines. LATA 2010 received 115 submissions, many among them of good quality. Each one was reviewed by at least three Program Committee members plus, in mostcases, byadditionalexternalreferees. Afterathoroughandvividdiscussion phase, the committee decided to accept 47 papers (which means an acceptance rate of 40. 86%). The conference program also included four invited talk
This book constitutes the thoroughly refereed post-conference proceedings of the Second IFIP TC 2 Central and East-European Conference on Software Engineering Techniques, CEE-SET 2008, held in Brno, Czech Republic, in October 2008. The 20 revised full papers presented together with a keynote speech were carefully reviewed and selected from 69 initial submissions. The papers are organized in topical sections on requirements specification, design, modeling, software product lines, code generation, project management, and quality.
The 7th Annual Conference on Theory and Applications of Models of Compu- tion was held during June 7-11, 2010 in Prague. After six successful conferences held in 2004-2009 in China (Beijing, Kunming, Beijing, Shanghai, Xi'an, and ChangSha) TAMC left Asia for the ?rst time, aiming at the "heart of Europe. " Changingthegeographicallocationdidnotbringanychangestothescopeofthe conference. Its three main themes continued to be Computability, Complexity, and Algorithms. The conference aims to bring together researchersfrom all over the world with interests in theoretical computer science, algorithmic mathem- ics,andapplicationsto thephysicalsciences. Thisyearwesawmoreparticipants from Europe and the Americas, but we were very happy that we could also w- come to Prague traditional participants from Asia (China, Japan, and India) to continue enhancing the collaboration among the theoretical computer science communities of these continents. After hard work the Program Committee decided to accept 35 papers out of 76 submitted to TAMC 2010. Each submission was reviewed by at least three, Program Committee members. All actions of the Program Committee were - ordinated via ?awlessly and e?ciently running EasyChair. We congratulate the authors of accepted contributions and thank all authors who submitted their papers. They all contributed to a successful event. We extend our thanks to the distinguished plenary speakers who accepted our invitation to deliver plenary talks - John Hopcroft from Cornell University and Shang-Hua Teng from University of Southern California. Their talks "New Research Directions in the Information Age" and "The Laplacian Paradigm: Emerging Algorithms for Massive Graph" were highlights of the conference.
When I attended college we studied vacuum tubes in our junior year. At that time an average radio had ?ve vacuum tubes and better ones even seven. Then transistors appeared in 1960s. A good radio was judged to be one with more thententransistors. Latergoodradioshad15-20transistors and after that everyone stopped counting transistors. Today modern processors runing personal computers have over 10milliontransistorsandmoremillionswillbeaddedevery year. The difference between 20 and 20M is in complexity, methodology and business models. Designs with 20 tr- sistors are easily generated by design engineers without any tools, whilst designs with 20M transistors can not be done by humans in reasonable time without the help of Prof. Dr. Gajski demonstrates the Y-chart automation. This difference in complexity introduced a paradigm shift which required sophisticated methods and tools, and introduced design automation into design practice. By the decomposition of the design process into many tasks and abstraction levels the methodology of designing chips or systems has also evolved. Similarly, the business model has changed from vertical integration, in which one company did all the tasks from product speci?cation to manufacturing, to globally distributed, client server production in which most of the design and manufacturing tasks are outsourced.
This book constitutes the refereed proceedings of the 4th International Conference, ICMT 2011, held in Zurich, Switzerland in June 2011. The 14 revised full papers were carefully revised and selected from 51 submissions. The scope of the contributions ranges from theoretical and methodological topics to implementation issues and applications. Topics addressed are such as transformation paradigms and languages, transformation algorithms and strategies, implementation and tools, as well as applications and case studies.
ETAPS 2010 was the 13th instance of the European Joint Conferences on T- oryandPracticeofSoftware. ETAPSisanannualfederatedconferencethatwas establishedin1998bycombininganumberofexistingandnewconferences. This yearitcomprisedtheusual?vesisterconferences(CC, ESOP, FASE, FOSSACS, TACAS), 19 satellite workshops (ACCAT, ARSPA-WITS, Bytecode, CMCS, COCV, DCC, DICE, FBTC, FESCA, FOSS-AMA, GaLoP, GT-VMT, LDTA, MBT, PLACES, QAPL, SafeCert, WGT, and WRLA) and seven invited l- tures (excluding those that were speci?c to the satellite events). The ?ve main conferences this year received 497 submissions (including 31 tool demonstration papers), 130 of which were accepted (10 tool demos), giving an overall acc- tance rateof 26%, with most of the conferencesat around24%. Congratulations thereforetoalltheauthorswhomadeittothe?nalprogramme Ihopethatmost of the other authors will still have found a way of participating in this exciting event, and that you will all continue submitting to ETAPS and contributing to make of it the best conference on software science and engineering. The events that comprise ETAPS address various aspects of the system - velopmentprocess, including speci?cation, design, implementation, analysisand improvement. The languages, methodologies and tools which support these - tivities are all well within its scope. Di?erent blends of theory and practice are represented, withaninclinationtowardtheorywithapracticalmotivationonthe one hand and soundly based practice on the other. Many of the issues involved in software design apply to systems in general, including hardwaresystems, and the emphasis on software is not intended to be exclusive. ETAPS is a confederation in which each event retains its own identity, with a separate Programme Committee and proceedings. Its format is open-ended, allowing it to grow and evolve as time go
ETAPS 2010 was the 13th instance of the European Joint Conferences on T- oryand Practiceof Software. ETAPS is anannual federatedconference that was establishedin1998bycombininganumberofexistingandnewconferences. This yearitcomprisedtheusual?vesisterconferences(CC,ESOP,FASE,FOSSACS, TACAS), 19 satellite workshops (ACCAT, ARSPA-WITS, Bytecode, CMCS, COCV, DCC, DICE, FBTC, FESCA, FOSS-AMA, GaLoP, GT-VMT, LDTA, MBT, PLACES, QAPL, SafeCert, WGT, and WRLA) and seven invited l- tures (excluding those that were speci?c to the satellite events). The ?ve main conferences this year received 497 submissions (including 31 tool demonstration papers), 130 of which were accepted (10 tool demos), giving an overall acc- tance rate of 26%, with most of the conferences at around 24%. Congratulations thereforetoalltheauthorswhomadeittothe?nalprogramme!Ihopethatmost of the other authors will still have found a way of participating in this exciting event, and that you will all continue submitting to ETAPS and contributing to make of it the best conference on software science and engineering. The events that comprise ETAPS address various aspects of the system - velopment process, including speci?cation, design, implementation, analysis and improvement. The languages, methodologies and tools which support these - tivities are all well within its scope. Di?erent blends of theory and practice are represented,withaninclinationtowardtheorywithapracticalmotivationonthe one hand and soundly based practice on the other. Many of the issues involved in software design apply to systems in general, including hardware systems, and the emphasis on software is not intended to be exclusive. ETAPS is a confederation in which each event retains its own identity, with a separate Programme Committee and proceedings.
ETAPS 2010 was the 13th instance of the European Joint Conferences on T- oryand Practiceof Software. ETAPS is anannual federatedconference that was establishedin1998bycombininganumberofexistingandnewconferences. This yearitcomprisedtheusual?vesisterconferences(CC, ESOP, FASE, FOSSACS, TACAS), 19 satellite workshops (ACCAT, ARSPA-WITS, Bytecode, CMCS, COCV, DCC, DICE, FBTC, FESCA, FOSS-AMA, GaLoP, GT-VMT, LDTA, MBT, PLACES, QAPL, SafeCert, WGT, and WRLA) and seven invited l- tures (excluding those that were speci?c to the satellite events). The ?ve main conferences this year received 497 submissions (including 31 tool demonstration papers), 130 of which were accepted (10 tool demos), giving an overall acc- tance rate of 26%, with most of the conferences at around 24%. Congratulations thereforetoalltheauthorswhomadeittothe?nalprogramme Ihopethatmost of the other authors will still have found a way of participating in this exciting event, and that you will all continue submitting to ETAPS and contributing to make of it the best conference on software science and engineering. The events that comprise ETAPS address various aspects of the system - velopment process, including speci?cation, design, implementation, analysis and improvement. The languages, methodologies and tools which support these - tivities are all well within its scope. Di?erent blends of theory and practice are represented, withaninclinationtowardtheorywithapracticalmotivationonthe one hand and soundly based practice on the other. Many of the issues involved in software design apply to systems in general, including hardware systems, and the emphasis on software is not intended to be exclusive. ETAPS is a confederation in which each event retains its own identity, with a separate Programme Committee and proceedi
Galois connections provide the order- or structure-preserving passage between two worlds of our imagination - and thus are inherent in hu man thinking wherever logical or mathematical reasoning about cer tain hierarchical structures is involved. Order-theoretically, a Galois connection is given simply by two opposite order-inverting (or order preserving) maps whose composition yields two closure operations (or one closure and one kernel operation in the order-preserving case). Thus, the "hierarchies" in the two opposite worlds are reversed or transported when passing to the other world, and going forth and back becomes a stationary process when iterated. The advantage of such an "adjoint situation" is that information about objects and relationships in one of the two worlds may be used to gain new information about the other world, and vice versa. In classical Galois theory, for instance, properties of permutation groups are used to study field extensions. Or, in algebraic geometry, a good knowledge of polynomial rings gives insight into the structure of curves, surfaces and other algebraic vari eties, and conversely. Moreover, restriction to the "Galois-closed" or "Galois-open" objects (the fixed points of the composite maps) leads to a precise "duality between two maximal subworlds.""
This book constitutes the refereed proceedings of the 18th International SPIN workshop on Model Checking Software, SPIN 2011, held in Snowbird, UT, USA, in July 2011. The 10 revised full papers presented together with 2 tool demonstration papers and 1 invited contribution were carefully reviewed and selected from 29 submissions. The papers are organized in topical sections on abstractions and state-space reductions; search strategies; PROMELA encodings and extensions; and applications of model checking.
In the multi-agent systems area, linking theory to practical applications is still a fertile research topic. The aim of the workshop on Declarative Agent Languages and Technologies (DALT 2009), in its seventh edition this year, is to achieve this goal, which needs developing and using advanced declarative technologies and languages, particularly agent programming, communication languages, and reasoning and decision-making mechanisms. Developing these technologies is a particularly challenging issue from many perspectives: formal foundations, pr- ticalfeasibility, degreeof?exibility, etc. Inthiscontext, thedeclarativeparadigm is arguably the most appropriate as unlike imperative approaches, the focus is onwhatthe solutionshouldaccomplishratherthanondescribing howto acc- plish it. This is because agentcomputing, as a paradigm, is about describing the logic of computation instead of describing how to accomplish it. DALT is about investigating, studying, andusing the declarativeparadigmaswell ascombining declarative and formal approaches with engineering and technology aspects of agents and multi-agent systems. This volume presents the latest developments in the area of declarative l- guagesandtechnologies, whichaimtoproviderigorousframeworksfordesigning, specifying, implementing and verifying autonomous interacting agents. These frameworksarebasedoncomputationallogicsand other formalmethods suchas mathematical models and game theoretical approaches. Using such models and approaches facilitates the development of agents that reason and act rationally while at the same time being able to verify the behavior of these agents against their speci?cation. The main theme of DALT 2009 was the further advan- ment of relevant speci?cation and veri?cation techniques, such as, for instance, modal and epistemic logics, model checking, constraint logic programming, and distributed constraint satisfa
This is the first textbook treatment of the algebraic approach to graph transformation, based on algebraic structures and category theory. It contains an introduction to classical graphs. Basic and advanced results are first shown for an abstract form of replacement systems and are then instantiated to several forms of graph and Petri net transformation systems. The book develops typed attributed graph transformation and contains a practical case study. |
You may like...
Programming Essentials Using Java
William McAllister
Mixed media product
UML 2002 - The Unified Modeling…
Jean-Marc Jezequel, Heinrich Hussman, …
Paperback
R1,586
Discovery Miles 15 860
Programming Finite Elements in Java (TM)
Gennadiy P. Nikishkov
Paperback
R2,438
Discovery Miles 24 380
Data Abstraction and Problem Solving…
Janet Prichard, Frank Carrano
Paperback
R2,280
Discovery Miles 22 800
Object-Oriented Analysis and Design for…
Raul Sidnei Wazlawick
Paperback
R1,120
Discovery Miles 11 200
|