![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Computer programming
Templates are used to generate all kinds of text, including computer code. The last decade, the use of templates gained a lot of popularity due to the increase of dynamic web applications. Templates are a tool for programmers, and implementations of template engines are most times based on practical experience rather than based on a theoretical background. This book reveals the mathematical background of templates and shows interesting findings for improving the practical use of templates. First, a framework to determine the necessary computational power for the template metalanguage is presented. The template metalanguage does not need to be Turing-complete to be useful. A non-Turing-complete metalanguage enforces separation of concerns between the view and model. Second, syntactical correctness of all languages of the templates and generated code is ensured. This includes the syntactical correctness of the template metalanguage and the output language. Third, case studies show that the achieved goals are applicable in practice. It is even shown that syntactical correctness helps to prevent cross-site scripting attacks in web applications. The target audience of this book is twofold. The first group exists of researcher interested in the mathematical background of templates. The second group exists of users of templates. This includes designers of template engines on one side and programmers and web designers using templates on the other side"
This book provides a comprehensive and practically minded introduction into serious games for law enforcement agencies. Serious games offer wide ranging benefits for law enforcement with applications from professional trainings to command-level decision making to the preparation for crises events. This book explains the conceptual foundations of virtual and augmented reality, gamification and simulation. It further offers practical guidance on the process of serious games development from user requirements elicitation to evaluation. The chapters are intended to provide principles, as well as hands-on knowledge to plan, design, test and apply serious games successfully in a law enforcement environment. A diverse set of case studies showcases the enormous variety that is possible in serious game designs and application areas and offers insights into concrete design decisions, design processes, benefits and challenges. The book is meant for law enforcement professionals interested in commissioning their own serious games as well as game designers interested in collaborative pedagogy and serious games for the law enforcement and security sector.
Bioinformatics is growing by leaps and bounds; theories/algorithms/statistical techniques are constantly evolving. Nevertheless, a core body of algorithmic ideas have emerged and researchers are beginning to adopt a "problem solving" approach to bioinformatics, wherein they use solutions to well-abstracted problems as building blocks to solve larger scope problems. "Problem Solving Handbook for Computational Biology" and Bioinformatics is an edited volume contributed by world renowned leaders in this field. This comprehensive handbook with problem solving emphasis, covers all relevant areas of computational biology and bioinformatics. Web resources and related themes are highlighted at every opportunity in this central easy-to-read reference. Designed for advanced-level students, researchers and professors in computer science and bioengineering as a reference or secondary text, this handbook is also suitable for professionals working in this industry.
Everything should be made as simple as possible, but not simpler. (Albert Einstein, Readers Digest, 1977) The modern practice of creating technical systems and technological processes of high effi.ciency besides the employment of new principles, new materials, new physical effects and other new solutions ( which is very traditional and plays the key role in the selection of the general structure of the object to be designed) also includes the choice of the best combination for the set of parameters (geometrical sizes, electrical and strength characteristics, etc.) concretizing this general structure, because the Variation of these parameters ( with the structure or linkage being already set defined) can essentially affect the objective performance indexes. The mathematical tools for choosing these best combinations are exactly what is this book about. With the advent of computers and the computer-aided design the pro bations of the selected variants are usually performed not for the real examples ( this may require some very expensive building of sample op tions and of the special installations to test them ), but by the analysis of the corresponding mathematical models. The sophistication of the mathematical models for the objects to be designed, which is the natu ral consequence of the raising complexity of these objects, greatly com plicates the objective performance analysis. Today, the main (and very often the only) available instrument for such an analysis is computer aided simulation of an object's behavior, based on numerical experiments with its mathematical model."
This book is an up-to-date self-contained compendium of the research carried out by the authors on model-based diagnosis of a class of discrete-event systems called active systems. After defining the diagnosis problem, the book copes with a variety of reasoning mechanisms that generate the diagnosis, possibly within a monitoring setting. The book is structured into twelve chapters, each of which has its own introduction and concludes with bibliographic notes and itemized summaries. Concepts and techniques are presented with the help of numerous examples, figures, and tables, and when appropriate these concepts are formalized into propositions and theorems, while detailed algorithms are expressed in pseudocode. This work is primarily intended for researchers, professionals, and graduate students in the fields of artificial intelligence and control theory.
This 2nd edition has been completely revised and updated, with additional new chapters. It presents state-of-the-art research in this area and focuses on key topics such as: visualization of semantic and structural information and metadata in the context of the emerging Semantic Web; Ontology-based Information Visualization and the use of graphically represented ontologies; Semantic Visualizations using Topic Maps and graph techniques; Recommender systems for filtering and recommending on the Semantic Web; SVG and X3D as new XML-based languages for 2D and 3D visualisations; methods used to construct and visualize high quality metadata and ontologies; and navigating and exploring XML documents using interactive multimedia interfaces. The design of visual interfaces for e-commerce and information retrieval is currently a challenging area of practical web development.
This monograph book is focused on the recent advances in smart, multimedia and computer gaming technologies. The Contributions include: *Smart Gamification and Smart Serious Games. *Fusion of secure IPsec-based Virtual Private Network, mobile computing and rich multimedia technology. *Teaching and Promoting Smart Internet of Things Solutions Using the Serious-game Approach. *Evaluation of Student Knowledge using an e-Learning Framework. *The iTEC Eduteka. *3D Virtual Worlds as a Fusion of Immersing, Visualizing, Recording, and Replaying Technologies. *Fusion of multimedia and mobile technology in audio guides for Museums and Exhibitions: from Bluetooth Push to Web Pull. The book is directed to researchers, students and software developers working in the areas of education and information technologies.
The implementation of object-oriented languages has been an active topic of research since the 1960s when the first Simula compiler was written. The topic received renewed interest in the early 1980s with the growing popularity of object-oriented programming languages such as c++ and Smalltalk, and got another boost with the advent of Java. Polymorphic calls are at the heart of object-oriented languages, and even the first implementation of Simula-67 contained their classic implementation via virtual function tables. In fact, virtual function tables predate even Simula-for example, Ivan Sutherland's Sketchpad drawing editor employed very similar structures in 1960. Similarly, during the 1970s and 1980s the implementers of Smalltalk systems spent considerable efforts on implementing polymorphic calls for this dynamically typed language where virtual function tables could not be used. Given this long history of research into the implementation of polymorphic calls, and the relatively mature standing it achieved over time, why, one might ask, should there be a new book in this field? The answer is simple. Both software and hardware have changed considerably in recent years, to the point where many assumptions underlying the original work in this field are no longer true. In particular, virtual function tables are no longer sufficient to implement polymorphic calls even for statically typed languages; for example, Java's interface calls cannot be implemented this way. Furthermore, today's processors are deeply pipelined and can execute instructions out-of order, making it difficult to predict the execution time of even simple code sequences."
This book delivers the latest developments in object technology and
their impact in computing systems re-engineering. Object-oriented
programming is here shown to provide support for constructing large
scale systems that are cheaply built and with reusable components,
adaptable to changing requirements and use efficient and
cost-effective techniques.
This new text/reference presents an accessible, concise, but rather complete, introduction to the C++ programming language with special emphasis on object-oriented numeric computation for scientific and engineering program development. The description of the language is in compliance with ISO/ANSI standards and is platform independent for maximum versatility. Requiring only basic calculus and linear algebra as prerequisites, the book introduces concepts, techniques, and standard libraries of C++ in a manner that is easy to understand and uses such familiar examples as vectors, matrices, integrals, and complex numbers. It also contains an introduction to C++ programs for applications with many numberic methods that are fundamental to science and engineering computing: polynomial evaluation and interpolation; numeric integration; methods for solving nonlinear equations; systems of linear equations in full, band, and sparse matrix storage formats; and ordinary and partial differential equations. Numerous techniques and examples are provided on how to reduce (C and Fortran) run-time overhead and improve program efficiency. Topics and features: *concise coverage of C++ programming concepts with object-oriented emphasis*numerous examples, coding tools, sample programs and exercises for reinforcement and self-study purposes*develops and uses basic and advanced features, as well as standard libraries of C++*covers many fundamental numeric methods for scientific and engineering computing applications*downloadable user-defined numeric linear algebra library available from author web site With an accessible style, intuitive topic development, and numerous examples, the book is an excellent resource and guide to the power, versatility and efficiency of C++ programming for numeric computing applications. Advanced students, practitioners and professionals in computer science, engineering and scientific computing in general will find the book a practical guide and resource for their work and applications program development.
This book introduces a fairly universal approach to the design and analysis of exact optimization algorithms for multi-objective combinatorial optimization problems. It proposes the circuits without repetitions representing the sets of feasible solutions along with the increasing and strictly increasing cost functions as a model for such problems. The book designs the algorithms for multi-stage and bi-criteria optimization and for counting the solutions in the framework of this model. As applications, this book studies eleven known combinatorial optimization problems: matrix chain multiplication, global sequence alignment, optimal paths in directed graphs, binary search trees, convex polygon triangulation, line breaking (text justification), one-dimensional clustering, optimal bitonic tour, segmented least squares, optimization of matchings in trees, and 0/1 knapsack problem. The results presented are useful for researchers in combinatorial optimization. This book is also useful as the basis for graduate courses.
A presentation of the central and basic concepts, techniques, and tools of computer science, with the emphasis on presenting a problem-solving approach and on providing a survey of all of the most important topics covered in degree programmes. Scheme is used throughout as the programming language and the author stresses a functional programming approach to create simple functions so as to obtain the desired programming goal. Such simple functions are easily tested individually, which greatly helps in producing programs that work correctly first time. Throughout, the author aids to writing programs, and makes liberal use of boxes with "Mistakes to Avoid." Programming examples include: * abstracting a problem; * creating pseudo code as an intermediate solution; * top-down and bottom-up design; * building procedural and data abstractions; * writing progams in modules which are easily testable. Numerous exercises help readers test their understanding of the material and develop ideas in greater depth, making this an ideal first course for all students coming to computer science for the first time.
Polynomial extremal problems (PEP) constitute one of the most important subclasses of nonlinear programming models. Their distinctive feature is that an objective function and constraints can be expressed by polynomial functions in one or several variables. Let: e = {: e 1, ...: en} be the vector in n-dimensional real linear space Rn; n PO(: e), PI (: e), ..., Pm (: e) are polynomial functions in R with real coefficients. In general, a PEP can be formulated in the following form: (0.1) find r = inf Po(: e) subject to constraints (0.2) Pi (: e) =0, i=l, ..., m (a constraint in the form of inequality can be written in the form of equality by introducing a new variable: for example, P( x) 0 is equivalent to P(: e) + y2 = 0). Boolean and mixed polynomial problems can be written in usual form by adding for each boolean variable z the equality: Z2 - Z = O. Let a = {al, ..., a } be integer vector with nonnegative entries {a;}f=l. n Denote by R a](: e) monomial in n variables of the form: n R a](: e) = IT: ef';;=1 d(a) = 2:7=1 ai is the total degree of monomial R a]. Each polynomial in n variables can be written as sum of monomials with nonzero coefficients: P(: e) = L caR a](: e), aEA{P) IX x Nondifferentiable optimization and polynomial problems where A(P) is the set of monomials contained in polynomial P
Overtheyears, research inthelifescienceshasbenefitedgreatlyfromthequantita tive toolsofmathematics and modeling. Many aspectsofcomplex biological systems can be more deeply understood when mathematical techniques are incorporated into a scientific investigation. Modelingcanbefruitfully applied in many typesofbiological research, from studies on the molecular, cellular, and organ level, to experiments in wholeanimalsandinpopulations. Using the field of nutrition as an example, one can find many cases of recent advances in knowledge and understanding that were facilitated by the application of mathematical modelingtokineticdata. Theavailabilityofbiologicallyimportantstable isotope-labeled compounds, developments in sensitive mass spectrometry and other analytical techniques, and advances in the powerful modeling software applied to data haveeachcontributed toourability tocarryoutevermoresophisticated kinetic studies that are relevant to nutrition and the health sciences at many levels oforganization. Furthermore, weanticipatethatmodeling isonthebrinkofanothermajoradvance: the application of kinetic modeling to clinical practice. With advances in the abilityof modelstoaccesslargedatabases(e. g., apopulationofindividualpatientrecords)andthe developmentofuserinterfaces thatare"friendly"enough tobeused byclinicians who arenotmodelers, wepredictthathealthapplicationsmodeling willbeanimportantnew 51 directionformodelinginthe21 century. This book contains manuscripts that are based on presentations at the seventh conference in a series focused on advancing nutrition and health research by fostering exchange among scientists from such disciplines as nutrition, biology, mathematics, statistics, kinetics, andcomputing. Thethemesofthesixpreviousconferencesincluded general nutritionmodeling(CanoltyandCain, 1985;Hoover-PlowandChandra, 1988), amino acids and carbohydrates (Aburnrad, 1991), minerals (Siva Subramanian and Wastney, 1995), vitamins, proteins, andmodelingtheory(CoburnandTownsend, 1996), and physiological compartmental modeling (Clifford and Muller, 1998). The seventh conference in the series was held at The Pennsylvania State University from July 29 throughAugust1,2000. Themeetingbeganwithaninstructiveandentertainingkeynote address by Professor Britton Chance, Eldridge Reeves Johnson University Professor Emeritus of Biophysics, Physical Chemistry, and Radiologic Physics, University of Pennsylvania. Dr."
Advice involves recommendations on what to think; through thought, on what to choose; and via choices, on how to act. Advice is information that moves by communication, from advisors to the recipient of advice. Ivan Jureta offers a general way to analyze advice. The analysis applies regardless of what the advice is about and from whom it comes or to whom it needs to be given, and it concentrates on the production and consumption of advice independent of the field of application. It is made up of two intertwined parts, a conceptual analysis and an analysis of the rationale of advice. He premises that giving advice is a design problem and he treats advice as an artifact designed and used to influence decisions. What is unusual is the theoretical backdrop against which the author's discussions are set: ontology engineering, conceptual analysis, and artificial intelligence. While classical decision theory would be expected to play a key role, this is not the case here for one principal reason: the difficulty of having relevant numerical, quantitative estimates of probability and utility in most practical situations. Instead conceptual models and mathematical logic are the author's tools of choice. The book is primarily intended for graduate students and researchers of management science. They are offered a general method of analysis that applies to giving and receiving advice when the decision problems are not well structured, and when there is imprecise, unclear, incomplete, or conflicting qualitative information.
Actor-Network Theory (ANT) has existed as a topic of interest among social theorists for decades. Due to the prevalence of technology in modern society, discussions over the influence of actor-network theory on the changing scope of technology can assist in facilitating further research and scientific thought. Technological Advancements and the Impact of Actor-Network Theory focuses on cross-disciplinary research as well as examples of the use of actor-network theory in a variety of fields, including medicine, education, business, engineering, environmental science, computer science, and social science. This timely publication is well-suited for reference use by academicians, researchers, upper-level students, and theorists. Topics Covered The many academic areas covered in this publication include, but are not limited to: Digital Communication E-Health Human Interaction Information and Communication Technologies Online Education Online Investing Public Service Innovation Software Development
This work introduces the benefits of object-oriented programming and discusses how the technology can be used to improve productivity in building software systems in the manufacturing domain. It addresses a wide range of issues from languages, design principles, research examples through to industrial applications and management issues. In essence, the main objective of the book is to interpret and apply object-oriented concepts in the context of designing manufacturing systems applications. The main audience for this book consists of professionals, engineers and managers, who deal with manufacturing systems, as well as students and educators looking for new directions in building software systems to solve problems in this area. The book should also be of special interest to engineering and computer professionals who have heard the term "object-oriented" and want to learn more about it and its importance, especially in designing software for manufacturing systems. This book should be of interest to: software and manufacturing engineers in industry; software consultants; technical managers; graduate students and researchers in computer-integrated manufacturing.
As the telecommunication industry introduces new sophisticated technologies, the nature of services and the volume of demands have changed. Indeed, a broad range of new services for users appear, combining voice, data, graphics, video, etc. This implies new planning issues. Fiber transmission systems that can carry large amounts of data on a few strands of wire were introduced. These systems have such a large bandwidth that the failure of even a single transmission link: in the network can create a severe service loss to customers. Therefore, a very high level of service reliability is becoming imperative for both system users and service providers. Since equipment failures and accidents cannot be avoided entirely, networks have to be designed so as to "survive" failures. This is done by judiciously installing spare capacity over the network so that all traffic interrupted by a failure may be diverted around that failure by way of this spare or reserve capacity. This of course translates into huge investments for network operators. Designing such survivable networks while minimizing spare capacity costs is, not surprisingly, a major concern of operating companies which gives rise to very difficult combinatorial problems. In order to make telecommunication networks survivable, one can essentially use two different strategies: protection or restoration. The protection approach preas signs spare capacity to protect each element of the network independently, while the restoration approach spreads the redundant capacity over the whole network and uses it as required in order to restore the disrupted traffic."
2 Concept ( Tools * Specification ( Tools + Design Stages ( Tools * Implementation ( Tools Figure 1-1. A nominal, multi-stage development process From that beginning, we have progressed to the point where the EDA community at large, including both users and developers of the tools, are interested in more unified environments. Here, the notion is that the tools used at the various stages in the development process need to be able to complement each other, and to communicate with one another efficiently using effective file exchange capabilities. Furthermore, the idea of capturing all the tool support needed for an EDA development into a unified support environment is now becoming a reality. This reality is evidenced by some of the EDA suites we now see emerging, wherein several tool functions are integrated under a common graphical user interface (GUI), with supporting file exchange and libraries to enable all tool functions to operate effectively and synergistically. This concept, which we illustrate in Figure 1- 2, is the true future ofEDA.
Providing a wide variety of technologies for ensuring the safety and dependability of cyber-physical systems (CPS), this book offers a comprehensive introduction to the architecture-centric modeling, analysis, and verification of CPS. In particular, it focuses on model driven engineering methods including architecture description languages, virtual prototyping, and formal analysis methods. CPS are based on a new design paradigm intended to enable emerging software-intensive systems. Embedded computers and networks monitor and control the physical processes, usually with the help of feedback loops where physical processes affect computations and vice versa. The principal challenges in system design lie in this constant interaction of software, hardware and physics. Developing reliable CPS has become a critical issue for the industry and society, because many applications such as transportation, power distribution, medical equipment and tele-medicine are dependent on CPS. Safety and security requirements must be ensured by means of powerful validation tools. Satisfying such requirements, including quality of service, implies having formally proven the required properties of the system before it is deployed. The book is concerned with internationally standardized modeling languages such as AADL, SysML, and MARTE. As the effectiveness of the technologies is demonstrated with industrial sample cases from the automotive and aerospace sectors, links between the methods presented and industrial problems are clearly understandable. Each chapter is self-contained, addressing specific scientific or engineering problems, and identifying further issues. In closing, it includes perspectives on future directions in CPS design from an architecture analysis viewpoint.
Many approaches have been proposed to enhance software productivity and reliability. These approaches typically fall into three categories: the engineering approach, the formal approach, and the knowledge-based approach. The optimal gain in software productivity cannot be obtained if one relies on only one of these approaches. Thus, the integration of different approaches has also become a major area of research. No approach can be said to be perfect if it fails to satisfy the following two criteria. Firstly, a good approach should support the full life cycle of software development. Secondly, a good approach should support the development of large-scale software for real use in many application domains. Such an approach can be referred to as a five-in-one approach. The authors of this book have, for the past eight years, conducted research in knowledge-based software engineering, of which the final goal is to develop a paradigm for software engineering which not only integrates the three approaches mentioned above, but also fulfils the two criteria on which the five-in-one approach is based. Domain Modeling- Based Software Engineering: A Formal Approach explores the results of this research. Domain Modeling-Based Software Engineering: A Formal Approach will be useful to researchers of knowledge-based software engineering, students and instructors of computer science, and software engineers who are working on large-scale projects of software development and want to use knowledge-based development methods in their work.
Aligning an organization's goals and strategies requires specifying their rationales and connections so that the links are explicit and allow for analytic reasoning about what is successful and where improvement is necessary. This book provides guidance on how to achieve this alignment, how to monitor the success of goals and strategies and use measurement to recognize potential failures, and how to close alignment gaps. It uses the GQM+Strategies approach, which provides concepts and actionable steps for creating the link between goals and strategies across an organization and allows for measurement-based decision-making. After outlining the general motivation for organizational alignment through measurement, the GQM+Strategies approach is described concisely, with a focus on the basic model that is created and the process for creating and using this model. The recommended steps of all six phases of the process are then described in detail with the help of a comprehensive application example. Finally, the industrial challenges addressed by the method and cases of its application in industry are presented, and the relations to other approaches, such as Balanced Scorecard, are described. The book concludes with supplementary material, such as checklists and guidelines, to support the application of the method. This book is aimed at organization leaders, managers, decision makers, and other professionals interested in aligning their organization's goals and strategies and establishing an efficient strategic measurement program. It is also interesting for academic researchers looking for mechanisms to integrate their research results into organizational environments.
System Design: A Practical Guide with SpecC presents the system design flow following a simple example through the whole process in an easy-to-follow, step-by-step fashion. Each step is described in detail in pictorial form and with code examples in SpecC. For each picture slide a detailed explanation is provided of the concepts presented. This format is suited for tutorials, seminars, self-study, as a guided reference carried by examples, or as teaching material for courses on system design. Features: Comprehensive introduction to and description of the SpecC language and design methodology; IP-centric language and methodology with focus on design reuse; Complete framework for system-level design from specification to implementation for SOCs and other embedded HW/SW systems. System Design: A Practical Guide with SpecC will benefit designers and design managers of complex SOCs, or embedded systems in general, by allowing them to develop new methodologies from these results, in order to increase design productivity by orders of magnitude. Designers at RTL, logical or physical levels, who are interested in moving up to the system level, will find a comprehensive overview within. The design models in the book define IP models and functions for IP exchange between IP providers and their users. A well-defined methodology like the one presented in this book will help product planning divisions to quickly develop new products or to derive completely new business models, like e-design or product-on-demand. Finally, researchers and students in the area of system design will find an example of a formal, well-structured design flow in this book. |
![]() ![]() You may like...
Dark Silicon and Future On-chip Systems…
Suyel Namasudra, Hamid Sarbazi-Azad
Hardcover
R4,186
Discovery Miles 41 860
Proceedings of International Conference…
Manish Prateek, T.P. Singh, …
Hardcover
R8,468
Discovery Miles 84 680
Big Data Analytics and Artificial…
Aboul Ella Hassanien, Nilanjan Dey, …
Hardcover
R2,911
Discovery Miles 29 110
C++ How to Program: Horizon Edition
Harvey Deitel, Paul Deitel
Paperback
R1,917
Discovery Miles 19 170
Flight Of The Diamond Smugglers - A Tale…
Matthew Gavin Frank
Paperback
|