![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming
Visual Knowledge Modeling for Semantic Web Technologies: Models and Ontologies aims to make visual knowledge modeling available to individuals as an intellectual method and a set of tools at different levels of formalization. It aims to provide to its readers a simple, yet powerful visual language to structure their thoughts, analyze information, transform it to personal knowledge, and communicate information to support knowledge acquisition in collaborative activities.
In recent years, cloud computing has gained a significant amount of attention by providing more flexible ways to store applications remotely. With software testing continuing to be an important part of the software engineering life cycle, the emergence of software testing in the cloud has the potential to change the way software testing is performed. Software Testing in the Cloud: Perspectives on an Emerging Discipline is a comprehensive collection of research by leading experts in the field providing an overview of cloud computing and current issues in software testing and system migration. Deserving the attention of researchers, practitioners, and managers, this book aims to raise awareness about this new field of study.
Through a series of step-by-step tutorials and numerous hands-on exercises, this book aims to equip the reader with both a good understanding of the importance of space in the abstract world of engineers and the ability to create a model of a product in virtual space - a skill essential for any designer or engineer who needs to present ideas concerning a particular product within a professional environment. The exercises progress logically from the simple to the more complex; while Solid Works or NX is the software used, the underlying philosophy is applicable to all modeling software. In each case, the explanation covers the entire procedure from the basic idea and production capabilities through to the real model; the conversion from 3D model to 2D manufacturing drawing is also clearly explained. Topics covered include modeling of prism, axisymmetric, symmetric and sophisticated shapes; digitization of physical models using modeling software; creation of a CAD model starting from a physical model; free form surface modeling; modeling of product assemblies following bottom-up and top-down principles; and the presentation of a product in accordance with the rules of technical documentation. This book, which includes more than 500 figures, will be ideal for students wishing to gain a sound grasp of space modeling techniques. Academics and professionals will find it to be an excellent teaching and research aid, and an easy-to-use guide.
This book offers a coherent and comprehensive approach to feature subset selection in the scope of classification problems, explaining the foundations, real application problems and the challenges of feature selection for high-dimensional data. The authors first focus on the analysis and synthesis of feature selection algorithms, presenting a comprehensive review of basic concepts and experimental results of the most well-known algorithms. They then address different real scenarios with high-dimensional data, showing the use of feature selection algorithms in different contexts with different requirements and information: microarray data, intrusion detection, tear film lipid layer classification and cost-based features. The book then delves into the scenario of big dimension, paying attention to important problems under high-dimensional spaces, such as scalability, distributed processing and real-time processing, scenarios that open up new and interesting challenges for researchers. The book is useful for practitioners, researchers and graduate students in the areas of machine learning and data mining.
The emphasis on new and changing technologies and process models in today s software development obscures the fact that software engineering is still primarily a human-based activity and that the success of a software project largely depends on the decisions made by humans during engineering. Rationale management is concerned with making these design and development decisions explicit to all stakeholders involved. Dutoit, McCall, Mistrik and Paech begin their book with a historical survey of different rationale approaches. It is followed by four parts describing: the fundamental problems and possible solution approaches in rationale management, rationale management during requirements engineering, rationale management during software architecting, and rationale management for organizing reusable bodies of knowledge. The result is a detailed summary of research on design rationale. It provides researchers with an excellent state-of-the-art overview, and professional software engineers will find many examples, resources and incentives to enhance their ability to make decisions during all phases of the software lifecycle. Allen Dutoit, Ray McCall, Ivan Mistrik and Barbara Paech have done an excellent job of this in "Rationale management in software engineering." The chapters in this volume show how design rationale can be incorporated into the heart of the software development process - into requirements engineering, software architecture, and code design. (John M. Carroll, Edward M. Frymoyer Professor of Information Sciences and Technology, Penn State University, USA, ACM CHI Lifetime Achievement Award)
The present book is the result of a three year research project which investigated the creative act of composing by means of algorithmic composition. Central to the investigation are the compositional strategies of 12 composers, which were documented through a dialogic and cyclic process of modelling and evaluating musical materials. The aesthetic premises and compositional approaches configure a rich spectrum of diverse positions, which is reflected also in the kinds of approaches and methods used. These approaches and methods include the generation and evaluation of chord sequences using genetic algorithms, the application of morphing strategies to research harmonic transformations, an automatic classification of personal preferences via machine learning, and an application of mathematical music theory to the analysis and resynthesis of musical material. The second part of the book features contributions by Sandeep Bhagwati, William Brooks, David Cope, Darla Crispin, Nicolas Donin, and Guerino Mazzola. These authors variously consider the project from different perspectives, offer independent approaches, or provide more general reflections from their respective research fields.
"Software Defined Networks" discusses the historical networking
environment that gave rise to SDN, as well as the latest advances
in SDN technology. The book gives you the state of the art
knowledge needed for successful deployment of an SDN, including:
How to explain to the non-technical business decision makers in
your organization the potential benefits, as well as the risks, in
shifting parts of a network to the SDN modelHow to make intelligent
decisions about when to integrate SDN technologies in a networkHow
to decide if your organization should be developing its own SDN
applications or looking to acquire these from an outside vendorHow
to accelerate the ability to develop your own SDN application, be
it entirely novel or a more efficient approach to a long-standing
problem
Model-driven software development drastically alters the software development process, which is characterized by a high degree of innovation and productivity. Emerging Technologies for the Evolution and Maintenance of Software Models contains original academic work about current research and research projects related to all aspects affecting the maintenance, evolution, and reengineering (MER), as well as long-term management, of software models. The mission of this book is to present a comprehensive and central overview of new and emerging trends in software model research and to provide concrete results from ongoing developments in the field.
NewInternetdevelopmentsposegreaterandgreaterprivacydilemmas. Inthe- formation Society, the need for individuals to protect their autonomy and retain control over their personal information is becoming more and more important. Today, informationandcommunicationtechnologies-andthepeopleresponsible for making decisions about them, designing, and implementing them-scarcely consider those requirements, thereby potentially putting individuals' privacy at risk. The increasingly collaborative character of the Internet enables anyone to compose services and contribute and distribute information. It may become hard for individuals to manage and control information that concerns them and particularly how to eliminate outdated or unwanted personal information, thus leavingpersonalhistoriesexposedpermanently. Theseactivitiesraisesubstantial new challenges for personal privacy at the technical, social, ethical, regulatory, and legal levels: How can privacy in emerging Internet applications such as c- laborative scenarios and virtual communities be protected? What frameworks and technical tools could be utilized to maintain life-long privacy? DuringSeptember3-10,2009, IFIP(InternationalFederationforInformation Processing)workinggroups9. 2 (Social Accountability),9. 6/11. 7(IT Misuseand theLaw),11. 4(NetworkSecurity)and11. 6(IdentityManagement)heldtheir5th InternationalSummerSchoolincooperationwiththeEUFP7integratedproject PrimeLife in Sophia Antipolis and Nice, France. The focus of the event was on privacy and identity managementfor emerging Internet applications throughout a person's lifetime. The aim of the IFIP Summer Schools has been to encourage young a- demic and industry entrants to share their own ideas about privacy and identity management and to build up collegial relationships with others. As such, the Summer Schools havebeen introducing participants to the social implications of information technology through the process of informed discussion.
Calculus has been used in solving many scientific and engineering problems. For optimization problems, however, the differential calculus technique sometimes has a drawback when the objective function is step-wise, discontinuous, or multi-modal, or when decision variables are discrete rather than continuous. Thus, researchers have recently turned their interests into metaheuristic algorithms that have been inspired by natural phenomena such as evolution, animal behavior, or metallic annealing. This book especially focuses on a music-inspired metaheuristic algorithm, harmony search. Interestingly, there exists an analogy between music and optimization: each musical instrument corresponds to each decision variable; musical note corresponds to variable value; and harmony corresponds to solution vector. Just like musicians in Jazz improvisation play notes randomly or based on experiences in order to find fantastic harmony, variables in the harmony search algorithm have random values or previously-memorized good values in order to find optimal solution.
This book presents an overview of the differential evolution algorithm. In the last few years the evolutionary computation domain has developed rapidly, and differential evolution is one of the representatives of this domain. It is a recently invented evolutionary algorithm that is gaining more and more popularity. Originally proposed for continuous unconstraint optimization, it was enlarged both for mixed optimization and for handling nonlinear constraints. Later on, new strategies, tuning, and adaptation of control parameters, ways of hybridization were elaborated. Attempts at theoretical analysis were accomplished as well. Moreover, the algorithm has a huge number of practical applications in different areas of science and industry.
Genetic algorithms today constitute a family of e?ective global optimization methods used to solve di?cult real-life problems which arise in science and technology. Despite their computational complexity, they have the ability to explore huge data sets and allow us to study exceptionally problematic cases in which the objective functions are irregular and multimodal, and where information about the extrema location is unobtainable in other ways. Theybelongtotheclassofiterativestochasticoptimizationstrategiesthat, during each step, produce and evaluate the set of admissible points from the search domain, called the random sample or population. As opposed to the Monte Carlo strategies, in which the population is sampled according to the uniform probability distribution over the search domain, genetic algorithms modify the probability distribution at each step. Mechanisms which adopt sampling probability distribution are transposed from biology. They are based mainly on genetic code mutation and crossover, as well as on selection among living individuals. Such mechanisms have been testedbysolvingmultimodalproblemsinnature,whichiscon?rmedinpart- ular by the many species of animals and plants that are well ?tted to di?erent ecological niches. They direct the search process, making it more e?ective than a completely random one (search with a uniform sampling distribution). Moreover,well-tunedgenetic-basedoperationsdonotdecreasetheexploration ability of the whole admissible set, which is vital in the global optimization process. The features described above allow us to regard genetic algorithms as a new class of arti?cial intelligence methods which introduce heuristics, well tested in other ?elds, to the classical scheme of stochastic global search.
This book gives a practical introduction to model-based testing,
showing how to write models for testing purposes and how to use
model-based testing tools to generate test suites. It is aimed at
testers and software developers who wish to use model-based
testing, rather than at tool-developers or academics.
Graphics and game developers must learn to program for mobility.
This book will teach you how.
The advent of powerful processing technologies and the advances in software development tools have drastically changed the approach and implementation of computational research in fundamental properties of living systems through simulating and synthesizing biological entities and processes in artificial media. Nowadays realistic physical and physiological simulation of natural and would-be creatures, worlds and societies becomes a low-cost task for ordinary home computers. The progress in technology has dramatically reshaped the structure of the software, the execution of a code, and visualization fundamentals. This has led to the emergence of novel breeds of artificial life software models, including three-dimensional programmable simulation environment, distributed discrete events platforms and multi-agent systems. This second edition reflects the technological and research advancements, and presents the best examples of artificial life software models developed in the World and available for users.
Over the next few years, the proprietary trading and hedge fund
industries will migrate largely to automated trade selection and
execution systems. Indeed, this is already happening. While several
finance books provide C++ code for pricing derivatives and
performing numerical calculations, none approaches the topic from a
system design perspective. This book will be divided into two
sections-programming techniques and automated trading system ( ATS
) technology-and teach financial system design and development from
the absolute ground up using Microsoft Visual C++.NET 2005. MS
Visual C++.NET 2005 has been chosen as the implementation language
primarily because most trading firms and large banks have developed
and continue to develop their proprietary algorithms in ISO C++ and
Visual C++.NET provides the greatest flexibility for incorporating
these legacy algorithms into working systems. Furthermore, the .NET
Framework and development environment provide the best libraries
and tools for rapid development of trading systems.
This book explores how agile development practices, in particular pair programming, code review and automated testing, help software development teams to perform better. Agile software engineering has become the standard software development paradigm over the last decade, and the insights provided here are taken from a large-scale survey of 80 professional software development teams working at SAP SE in Germany. In addition, the book introduces a novel measurement tool for assessing the performance of software development teams. No previous study has researched this topic with a similar data set comprising insights from more than 450 professional software engineers.
In development scenarios where things can't be run in a debugger, or when you run the risk of masking the problem, logs are the greatest source of information about running a program. "Pro Apache Log4j, Second Edition" provides best practices guidelines and comprehensive coverage of the most recent release. Step by step, the book explains core concepts, from basic to advanced. Code samples are in Java and include guidelines for different application-specific needs. You'll also learn how to extend the API to write custom components and best practices for using the feature-rich log4j API. This book concludes with enterprise Java applications using log4j with JSP and J2EE.
"As projects get more complicated, managers stop learning from their - perience. It is important to understand how that happens and how to change it.... Fallible estimates: In software development, initial estimates for a project shape the trajectory of decisions that a manager makes over its life. For ex- ple, estimates of the productivity of the team members influence decisions about the size of the team, which in turn affect the team's actual output. The trouble is that initial estimates usually turn out to be wrong. " (Sengupta, 2008) This book aims directly to increase the awareness among managers and practitioners that estimation is as important as the work to be done in so- ware and systems development. You can manage what you can measure! Readers will find in this book a collection of lessons learned from the worldwide "metrics community," which we have documented and enhanced with our own experiences in the field of software measurement and estimating. Our goal is to support our readers to harvest the benefits of estimating and - prove their software development processes. We present the 5 ISO/I- acknowledged Functional Sizing Methods with variants, experiences, counting rules, and case studies - and most importantly, illustrate through practical - amples how to use functional size measurement to produce realistic estimates. The book is written in a practical manner, especially for the busy practitioner community. It is aimed to be used as a manual and an assistant for everyday work.
This book presents four mathematical essays which explore the foundations of mathematics and related topics ranging from philosophy and logic to modern computer mathematics. While connected to the historical evolution of these concepts, the essays place strong emphasis on developments still to come. The book originated in a 2002 symposium celebrating the work of Bruno Buchberger, Professor of Computer Mathematics at Johannes Kepler University, Linz, Austria, on the occasion of his 60th birthday. Among many other accomplishments, Professor Buchberger in 1985 was the founding editor of the Journal of Symbolic Computation; the founder of the Research Institute for Symbolic Computation (RISC) and its chairman from 1987-2000; the founder in 1990 of the Softwarepark Hagenberg, Austria, and since then its director. More than a decade in the making, Mathematics, Computer Science and Logic - A Never Ending Story includes essays by leading authorities, on such topics as mathematical foundations from the perspective of computer verification; a symbolic-computational philosophy and methodology for mathematics; the role of logic and algebra in software engineering; and new directions in the foundations of mathematics. These inspiring essays invite general, mathematically interested readers to share state-of-the-art ideas which advance the never ending story of mathematics, computer science and logic. Mathematics, Computer Science and Logic - A Never Ending Story is edited by Professor Peter Paule, Bruno Buchberger s successor as director of the Research Institute for Symbolic Computation. "
The book discusses intelligent system design using soft computing and similar systems and their interdisciplinary applications. It also focuses on the recent trends to use soft computing as a versatile tool for designing a host of decision support systems. |
You may like...
Intro to Python for Computer Science and…
Paul Deitel
Paperback
Flash Memory Integration - Performance…
Jalil Boukhobza, Pierre Olivier
Hardcover
R1,831
Discovery Miles 18 310
|