![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming
This monograph is concerned with the problem of getting computers to transform formal language definitions into compilers. Its purpose is to demonstrate how certain simple theoretical ideas can be used to generate compilers and even compiler generators. As the title suggests, a realistic assessment of the relationship between the complexity of realistic compilation and the relative simplicity studied in theoretical work is attempted. The monograph contains an overview of existing compiler generators. The CERES '83 compiler generator, developed by Neil D. Jones and the author, is described in detail. The CERES system is based on the idea of composing language definitions and it serves as an example of a powerful novel "bootstrapping" technique by which one can generate compiler generators as well as compilers by considering a compiler generator to be, in a sense which is made mathematically precise, a special kind of compiler. The core of the CERES system is a two-page-long machine generated compiler generator. The approach uses ideas from denotational semantics and many-sorted algebra and connects them with novel ideas about how to treat programs and language definitions as data. Considerable effort has been made to present the necessary theory in a manner suitable for readers who have some practical experience but not necessarily a theoretical background in semantics.
Algorithmic Principles of Mathematical Programming investigates the
mathematical structures and principles underlying the design of
efficient algorithms for optimization problems. Recent advances in
algorithmic theory have shown that the traditionally separate areas
of discrete optimization, linear programming, and nonlinear
optimization are closely linked. This book offers a comprehensive
introduction to the whole subject and leads the reader to the
frontiers of current research. The prerequisites to use the book
are very elementary. All the tools from numerical linear algebra
and calculus are fully reviewed and developed. Rather than
attempting to be encyclopedic, the book illustrates the important
basic techniques with typical problems. The focus is on efficient
algorithms with respect to practical usefulness. Algorithmic
complexity theory is presented with the goal of helping the reader
understand the concepts without having to become a theoretical
specialist. Further theory is outlined and supplemented with
pointers to the relevant literature.
This book is devoted to one of the main questions of the theory of extremal prob lems, namely, to necessary and sufficient extremality conditions. It is intended mostly for mathematicians and also for all those who are interested in optimiza tion problems. The book may be useful for advanced students, post-graduated students, and researchers. The book consists of four chapters. In Chap. 1 we study the abstract minimization problem with constraints, which is often called the mathemati cal programming problem. Chapter 2 is devoted to one of the most important classes of extremal problems, the optimal control problem. In the third chapter we study one of the main objects of the calculus of variations, the integral quadratic form. In the concluding, fourth, chapter we study local properties of smooth nonlinear mappings in a neighborhood of an abnormal point. The problems which are studied in this book (of course, in addition to their extremal nature) are united by our main interest being in the study of the so called abnormal or degenerate problems. This is the main distinction of the present book from a large number of books devoted to theory of extremal problems, among which there are many excellent textbooks, and books such as, e.g., 13, 38, 59, 78, 82, 86, 101, 112, 119], to mention a few."
also in: THE KLUWER INTERNATIONAL SERIES ON ASIAN STUDIES IN COMPUTER AND INFORMATION SCIENCE, Volume 1
The volume provides a comprehensive, up-to-date account on recent developments concerning the incorporation of fuzzy capabilities in Petri Net models. The results of such studies originated the class of models that have been designated by Fuzzy Petri Nets. The recent papers specially elaborated for this volume range over several aspects of fuzziness in Petri nets. They form an interesting collection of original works that covers a great variety of relevant problems concerning the concept of Fuzzy Petri Net model. The articles approach several of the most outstanding issues in the framework of Fuzzy Petri nets, such as the representation of time, consistency checking, learning, design, computational efficiency, modelling flexibility, among others. From the material collected in the book one can extract the points of view of leading researchers concerning the basic and advanced concepts, advantages, potential applications and open problems, related to the field.
This book discusses research, methods, and recent developments in the interdisciplinary field that spans research in visualization, eye tracking, human-computer interaction, and psychology. It presents extended versions of papers from the First Workshop on Eye Tracking and Visualization (ETVIS), which was organized as a workshop of the IEEE VIS Conference 2015. Topics include visualization and visual analytics of eye-tracking data, metrics and cognitive models, eye-tracking experiments in the context of visualization interfaces, and eye tracking in 3D and immersive environments. The extended ETVIS papers are complemented by a chapter offering an overview of visualization approaches for analyzing eye-tracking data and a chapter that discusses electrooculography (EOG) as an alternative of acquiring information about eye movements. Covering scientific visualization, information visualization, and visual analytics, this book is a valuable resource for eye-tracking researchers within the visualization community.
Articles in this book examine various materials and how to determine directly the limit state of a structure, in the sense of limit analysis and shakedown analysis. Apart from classical applications in mechanical and civil engineering contexts, the book reports on the emerging field of material design beyond the elastic limit, which has further industrial design and technological applications. Readers will discover that "Direct Methods" and the techniques presented here can in fact be used to numerically estimate the strength of structured materials such as composites or nano-materials, which represent fruitful fields of future applications. Leading researchers outline the latest computational tools and optimization techniques and explore the possibility of obtaining information on the limit state of a structure whose post-elastic loading path and constitutive behavior are not well defined or well known. Readers will discover how Direct Methods allow rapid and direct access to requested information in mathematically constructive manners without cumbersome step-by-step computation. Both researchers already interested or involved in the field and practical engineers who want to have a panorama of modern methods for structural safety assessment will find this book valuable. It provides the reader with the latest developments and a significant amount of references on the topic.
Distributed Infrastructure Support For E-Commerce And Distributed
Applications is organized in three parts. The first part
constitutes an overview, a more detailed motivation of the problem
context, and a tutorial-like introduction to middleware systems.
The second part is comprised of a set of chapters that study
solutions to leverage the trade-off between a transparent
programming model and application-level enabled resource control.
The third part of this book presents three detailed distributed
application case studies and demonstrates how standard middleware
platforms fail to adequately cope with resource control needs of
the application designer in these three cases:
Optimization in Computational Chemistry and Molecular Biology: Local and Global Approaches covers recent developments in optimization techniques for addressing several computational chemistry and biology problems. A tantalizing problem that cuts across the fields of computational chemistry, biology, medicine, engineering and applied mathematics is how proteins fold. Global and local optimization provide a systematic framework of conformational searches for the prediction of three-dimensional protein structures that represent the global minimum free energy, as well as low-energy biomolecular conformations. Each contribution in the book is essentially expository in nature, but of scholarly treatment. The topics covered include advances in local and global optimization approaches for molecular dynamics and modeling, distance geometry, protein folding, molecular structure refinement, protein and drug design, and molecular and peptide docking. Audience: The book is addressed not only to researchers in mathematical programming, but to all scientists in various disciplines who use optimization methods in solving problems in computational chemistry and biology.
Describes a small verification library with a concentration on user adaptability such as re-useable components, portable Intellectual Property, and co-verification. Includes a free CD of TEAL along with examples. Takes a realistic view of reusability and distills lessons learned down to a tool box of techniques and guidelines.
Among the most important problems confronting computer science is that of developing a paradigm appropriate to the discipline. Proponents of formal methods - such as John McCarthy, C.A.R. Hoare, and Edgar Dijkstra - have advanced the position that computing is a mathematical activity and that computer science should model itself after mathematics. Opponents of formal methods - by contrast, suggest that programming is the activity which is fundamental to computer science and that there are important differences that distinguish it from mathematics, which therefore cannot provide a suitable paradigm. Disagreement over the place of formal methods in computer science has recently arisen in the form of renewed interest in the nature and capacity of program verification as a method for establishing the reliability of software systems. A paper that appeared in Communications of the ACM entitled, Program Verification: The Very Idea', by James H. Fetzer triggered an extended debate that has been discussed in several journals and that has endured for several years, engaging the interest of computer scientists (both theoretical and applied) and of other thinkers from a wide range of backgrounds who want to understand computer science as a domain of inquiry. The editors of this collection have brought together many of the most interesting and important studies that contribute to answering questions about the nature and the limits of computer science. These include early papers advocating the mathematical paradigm by McCarthy, Naur, R. Floyd, and Hoare (in Part I), others that elaborate the paradigm by Hoare, Meyer, Naur, and Scherlis and Scott (in Part II), challenges, limits and alternatives explored by C. Floyd, Smith, Blum, and Naur (in Part III), and recent work focusing on formal verification by DeMillo, Lipton, and Perlis, Fetzer, Cohn, and Colburn (in Part IV). It provides essential resources for further study. This volume will appeal to scientists, philosophers, and laypersons who want to understand the theoretical foundations of computer science and be appropriately positioned to evaluate the scope and limits of the discipline.
The book offers an original view on channel coding, based on a unitary approach to block and convolutional codes for error correction. It presents both new concepts and new families of codes. For example, lengthened and modified lengthened cyclic codes are introduced as a bridge towards time-invariant convolutional codes and their extension to time-varying versions. The novel families of codes include turbo codes and low-density parity check (LDPC) codes, the features of which are justified from the structural properties of the component codes. Design procedures for regular LDPC codes are proposed, supported by the presented theory. Quasi-cyclic LDPC codes, in block or convolutional form, represent one of the most original contributions of the book. The use of more than 100 examples allows the reader gradually to gain an understanding of the theory, and the provision of a list of more than 150 definitions, indexed at the end of the book, permits rapid location of sought information.
Since the introduction of genetic algorithms in the 1970s, an enormous number of articles together with several significant monographs and books have been published on this methodology. As a result, genetic algorithms have made a major contribution to optimization, adaptation, and learning in a wide variety of unexpected fields. Over the years, many excellent books in genetic algorithm optimization have been published; however, they focus mainly on single-objective discrete or other hard optimization problems under certainty. There appears to be no book that is designed to present genetic algorithms for solving not only single-objective but also fuzzy and multiobjective optimization problems in a unified way. Genetic Algorithms And Fuzzy Multiobjective Optimization introduces the latest advances in the field of genetic algorithm optimization for 0-1 programming, integer programming, nonconvex programming, and job-shop scheduling problems under multiobjectiveness and fuzziness. In addition, the book treats a wide range of actual real world applications. The theoretical material and applications place special stress on interactive decision-making aspects of fuzzy multiobjective optimization for human-centered systems in most realistic situations when dealing with fuzziness. The intended readers of this book are senior undergraduate students, graduate students, researchers, and practitioners in the fields of operations research, computer science, industrial engineering, management science, systems engineering, and other engineering disciplines that deal with the subjects of multiobjective programming for discrete or other hard optimization problems under fuzziness. Real world research applications are used throughout the book to illustrate the presentation. These applications are drawn from complex problems. Examples include flexible scheduling in a machine center, operation planning of district heating and cooling plants, and coal purchase planning in an actual electric power plant.
The scientific monograph of a survey kind presented to the reader's attention deals with fundamental ideas and basic schemes of optimization methods that can be effectively used for solving strategic planning and operations manage ment problems related, in particular, to transportation. This monograph is an English translation of a considerable part of the author's book with a similar title that was published in Russian in 1992. The material of the monograph embraces methods of linear and nonlinear programming; nonsmooth and nonconvex optimization; integer programming, solving problems on graphs, and solving problems with mixed variables; rout ing, scheduling, solving network flow problems, and solving the transportation problem; stochastic programming, multicriteria optimization, game theory, and optimization on fuzzy sets and under fuzzy goals; optimal control of systems described by ordinary differential equations, partial differential equations, gen eralized differential equations (differential inclusions), and functional equations with a variable that can assume only discrete values; and some other methods that are based on or adjoin to the listed ones."
How to Write Code You're Proud of . . . Every Single Day ". . . [A] timely and humble reminder of the ever-increasing complexity of our programmatic world and how we owe it to the legacy of humankind--and to ourselves--to practice ethical development. Take your time reading Clean Craftsmanship. . . . Keep this book on your go-to bookshelf. Let this book be your old friend--your Uncle Bob, your guide--as you make your way through this world with curiosity and courage." --From the Foreword by Stacia Heimgartner Viscardi, CST & Agile Mentor In Clean Craftsmanship, the legendary Robert C. Martin ("Uncle Bob") has written the principles that define the profession--and the craft--of software development. Uncle Bob brings together the disciplines, standards, and ethics you need to deliver robust, effective code and to be proud of all the software you write. Robert Martin, the best-selling author of Clean Code, provides a pragmatic, technical, and prescriptive guide to the foundational disciplines of software craftsmanship. He discusses standards, showing how the world's expectations of developers often differ from their own and helping you bring the two in sync. Bob concludes with the ethics of the programming profession, describing the fundamental promises all developers should make to their colleagues, their users, and, above all, themselves. With Uncle Bob's insights, all programmers and their managers can consistently deliver code that builds trust instead of undermining it--trust among users and throughout societies that depend on software for their survival. Moving towards the "north star" of true software craftsmanship: the state of knowing how to program well Practical, specific guidance for applying five core disciplines: test-driven development, refactoring, simple design, collaborative programming, and acceptance tests How developers and teams can promote productivity, quality, and courage The true meaning of integrity and teamwork among programmers, and ten specific commitments every software professional should make Register your book for convenient access to the book's companion videos, updates, and/or corrections as they become available. See inside book for details.
Over the past several years, a great deal of research has been devoted to the use of information technology by small businesses. One technological tool now used to boost company success is Web presence enhancement in alignment with business strategy. ""Effective Web Presence Solutions for Small Businesses: Strategies for Successful Implementation"" is the first book to provide small businesses with a holistic approach to implementing their Web presence through identification of Web site content that matches their business strategy. A valuable read for small business owners as well as academicians and researchers, this book connects the various issues involved in the planning and execution of successful Web sites for small businesses.
Intuitionistic type theory can be described, somewhat boldly, as a partial fulfillment of the dream of a universal language for science. This book expounds several aspects of intuitionistic type theory, such as the notion of set, reference vs. computation, assumption, and substitution. Moreover, the book includes philosophically relevant sections on the principle of compositionality, lingua characteristica, epistemology, propositional logic, intuitionism, and the law of excluded middle. Ample historical references are given throughout the book.
How do you design personalized user experiences that delight and
provide value to the customers of an eCommerce site?
Personalization does not guarantee high quality user experience: a
personalized user experience has the best chance of success if it
is developed using a set of best practices in HCI. In this book 35
experts from academia, industry and government focus on issues in
the design of personalized web sites. The topics range from the
design and evaluation of user interfaces and tools to information
architecture and computer programming related to commercial web
sites. The book covers four main areas:
Since I started working in the area of nonlinear programming and, later on, variational inequality problems, I have frequently been surprised to find that many algorithms, however scattered in numerous journals, monographs and books, and described rather differently, are closely related to each other. This book is meant to help the reader understand and relate algorithms to each other in some intuitive fashion, and represents, in this respect, a consolidation of the field. The framework of algorithms presented in this book is called Cost Approxi mation. (The preface of the Ph.D. thesis Pat93d] explains the background to the work that lead to the thesis, and ultimately to this book.) It describes, for a given formulation of a variational inequality or nonlinear programming problem, an algorithm by means of approximating mappings and problems, a principle for the update of the iteration points, and a merit function which guides and monitors the convergence of the algorithm. One purpose of this book is to offer this framework as an intuitively appeal ing tool for describing an algorithm. One of the advantages of the framework, or any reasonable framework for that matter, is that two algorithms may be easily related and compared through its use. This framework is particular in that it covers a vast number of methods, while still being fairly detailed; the level of abstraction is in fact the same as that of the original problem statement."
The first textbook ever to cover multi-relational data mining and inductive logic programming, this book fully explores logical and relational learning. Ideal for graduate students and researchers, it also looks at statistical relational learning.
The book emphasizes the design of full-fledged, fully
normalizing lambda calculus
PHP is rapidly becoming the language of choice for dynamic Web development, in particular for e-commerce and on-line database systems. It is open source software and easy to install, and can be used with a variety of operating systems, including Microsoft Windows and UNIX. This comprehensive manual covers the basic core of the language, with lots of practical examples of some of the more recent and useful features available in version 5.0. MySQL database creation and development is also covered, as it is the developer database most commonly used alongside PHP. It will be an invaluable book for professionals wanting to use PHP to develop their own dynamic web pages. Key Topics: - Basic Language Constructs - Manipulating Arrays and Strings - Errors and Buffering - Graphic Manipulation - PDF Library Extension - MySQL Database Management - Classes and Objects Concepts Features and Benefits: Explains how to use PHP to its full extent - covering the latest features and functions of PHP version 5.0, including the use of object-oriented programming Describes how to link a database to a web site, using the MySQL database management system Shows how to connect PHP to other systems and provides many examples, so that you can create powerful and dynamic web pages and applications Contains lots of illustrated, practical, real-world examples - including an e-commerce application created in PHP using many of the features described within the book The scripts used in the examples are available for download from www.phpmysql-manual.com
As is true of most technological fields, the software industry is constantly advancing and becoming more accessible to a wider range of people. The advancement and accessibility of these systems creates a need for understanding and research into their development. Optimizing Contemporary Application and Processes in Open Source Software is a critical scholarly resource that examines the prevalence of open source software systems as well as the advancement and development of these systems. Featuring coverage on a wide range of topics such as machine learning, empirical software engineering and management, and open source, this book is geared toward academicians, practitioners, and researchers seeking current and relevant research on the advancement and prevalence of open source software systems.
Mixed-Signal Embedded Microcontrollers are commonly used in integrating analog components needed to control non-digital electronic systems. They are used in automatically controlled devices and products, such as automobile engine control systems, wireless remote controllers, office machines, home appliances, power tools, and toys. Microcontrollers make it economical to digitally control even more devices and processes by reducing the size and cost, compared to a design that uses a separate microprocessor, memory, and input/output devices. In many undergraduate and post-graduate courses, teaching of mixed-signal microcontrollers and their use for project work has become compulsory. Students face a lot of difficulties when they have to interface a microcontroller with the electronics they deal with. This book addresses some issues of interfacing the microcontrollers and describes some project implementations with the Silicon Lab C8051F020 mixed-signal microcontroller. The intended readers are college and university students specializing in electronics, computer systems engineering, electrical and electronics engineering; researchers involved with electronics based system, practitioners, technicians and in general anybody interested in microcontrollers based projects.
This book provides the most updated information of how membrane lipids mediate protein signaling from studies carried out in animal and plant cells. Also, there are some chapters that go beyond and expand these studies of protein-lipid interactions at the structural level. The book begins with a literature review from investigations associated to sphingolipids, followed by studies that describe the role of phosphoinositides in signaling and closing with the function of other key lipids in signaling at the plasma membrane and intracellular organelles. |
You may like...
The Testing and Learning Revolution…
Kavitha Rajagopalan, Edmund W. Gordon
Hardcover
R1,396
Discovery Miles 13 960
Introduction to Parallel Computing - A…
Wesley Petersen, Peter Arbenz
Hardcover
R5,836
Discovery Miles 58 360
Distributed and Parallel Systems - From…
Peter Kacsuk, Gabriele Kotsis
Hardcover
R5,273
Discovery Miles 52 730
Parallel Computing: Fundamentals…
E.H. D'Hollander, G.R. Joubert, …
Hardcover
R6,688
Discovery Miles 66 880
|