![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming
"Fixed-Point Algorithms for Inverse Problems in Science and Engineering" presents some of the most recent work from top-notch researchers studying projection and other first-order fixed-point algorithms in several areas of mathematics and the applied sciences. The material presented provides a survey of the state-of-the-art theory and practice in fixed-point algorithms, identifying emerging problems driven by applications, and discussing new approaches for solving these problems. This book incorporates diverse perspectives from broad-ranging areas of research including, variational analysis, numerical linear algebra, biotechnology, materials science, computational solid-state physics, and chemistry. Topics presented include: Theory of Fixed-point algorithms: convex analysis, convex optimization, subdifferential calculus, nonsmooth analysis, proximal point methods, projection methods, resolvent and related fixed-point theoretic methods, and monotone operator theory. Numerical analysis of fixed-point algorithms: choice of step lengths, of weights, of blocks for block-iterative and parallel methods, and of relaxation parameters; regularization of ill-posed problems; numerical comparison of various methods. Areas of Applications: engineering (image and signal reconstruction and decompression problems), computer tomography and radiation treatment planning (convex feasibility problems), astronomy (adaptive optics), crystallography (molecular structure reconstruction), computational chemistry (molecular structure simulation) and other areas. Because of the variety of applications presented, this book can easily serve as a basis for new and innovated research and collaboration.
Software effort estimation is a key element of software project planning and management. Yet, in industrial practice, the important role of effort estimation is often underestimated and/or misunderstood. In this book, Adam Trendowicz presents the CoBRA method (an abbreviation for Cost Estimation, Benchmarking, and Risk Assessment) for estimating the effort required to successfully complete a software development project, which uniquely combines human judgment and measurement data in order to systematically create a custom-specific effort estimation model. CoBRA goes far beyond simply predicting the development effort; it supports project decision-makers in negotiating the project scope, managing project risks, benchmarking productivity, and directing improvement activities. To illustrate the method's practical use, the book reports several real-world cases where CoBRA was applied in various industrial contexts. These cases represent different estimation contexts in terms of software project environment, estimation objectives, and estimation constraints. This book is the result of a successful collaboration between the process management division of Fraunhofer IESE and many software companies in the field of software engineering technology transfer. It mainly addresses software practitioners who deal with planning and managing software development projects as part of their daily work, and is also of interest for students or courses specializing in software engineering or software project management.
The past decades have seen significant improvements in 3D imaging where the related techniques and technologies have advanced to a mature state. These exciting developments have sparked increasing interest in the challenges and opportunities afforded by 3D sensing. As a consequence, the emerging area of safety and security related imaging incorporates these important new technologies beyond the limitations of 2D image processing.This book presents the thoroughly revised versions of lectures given by leading researchers during the Workshop on Advanced 3D Imaging for Safety and Security in conjunction with the International Conference on Computer Vision and Pattern Recognition CVPR 2005, held in San Diego, CA, USA in June 2005.It covers the current state of the art in 3D imaging for safety and security.
Graphs are widely used to represent structural information in the form of objects and connections between them. Graph transformation is the rule-based manipulation of graphs, an increasingly important concept in computer science and related fields. This is the first textbook treatment of the algebraic approach to graph transformation, based on algebraic structures and category theory. Part I is an introduction to the classical case of graph and typed graph transformation. In Part II basic and advanced results are first shown for an abstract form of replacement systems, so-called adhesive high-level replacement systems based on category theory, and are then instantiated to several forms of graph and Petri net transformation systems. Part III develops typed attributed graph transformation, a technique of key relevance in the modeling of visual languages and in model transformation. Part IV contains a practical case study on model transformation and a presentation of the AGG (attributed graph grammar) tool environment. Finally the appendix covers the basics of category theory, signatures and algebras. The book addresses both research scientists and graduate students in computer science, mathematics and engineering.
."NET 2.0 for Delphi Programmers" explores .NET from a Delphi programmers viewpoint, and it is ideal for Delphi programmers moving to .NET. It presents the core concepts of the .NET world in terms you are familiar with. This book will help you with Delphi for .NET as well as C#. Apress publishes migration books for both VB 6 and C++ programmers moving to .NET. Consider this the Delphi installment of Apress migration books! There is ample coverage of C# as well as Delphi for .NET inside this newest addition.
This book seeks to promote the structured, standardized and accurate use of software measurement at all levels of modern software development companies. To do so, it focuses on seven main aspects: sound scientific foundations, cost-efficiency, standardization, value-maximization, flexibility, combining organizational and technical aspects, and seamless technology integration. Further, it supports companies in their journey from manual reporting to automated decision support by combining academic research and industrial practice. When scientists and engineers measure something, they tend to focus on two different things. Scientists focus on the ability of the measurement to quantify whatever is being measured; engineers, however, focus on finding the right qualities of measurement given the designed system (e.g. correctness), the system's quality of use (e.g. ease of use), and the efficiency of the measurement process. In this book, the authors argue that both focuses are necessary, and that the two are complementary. Thus, the book is organized as a gradual progression from theories of measurement (yes, you need theories to be successful!) to practical, organizational aspects of maintaining measurement systems (yes, you need the practical side to understand how to be successful). The authors of this book come from academia and industry, where they worked together for the past twelve years. They have worked with both small and large software development organizations, as researchers and as measurement engineers, measurement program leaders and even teachers. They wrote this book to help readers define, implement, deploy and maintain company-wide measurement programs, which consist of a set of measures, indicators and roles that are built around the concept of measurement systems. Based on their experiences introducing over 40,000 measurement systems at over a dozen companies, they share essential tips and tricks on how to do it right and how to avoid common pitfalls.
2012 Jolt Award finalist! Pioneering the Future of Software Test Do you need to get it right, too? Then, learn from Google. Legendary testing expert James Whittaker, until recently a Google testing leader, and two top Google experts reveal exactly how Google tests software, offering brand-new best practices you can use even if you're not quite Google's size...yet! Breakthrough Techniques You Can Actually Use Discover 100% practical, amazingly scalable techniques for analyzing risk and planning tests...thinking like real users...implementing exploratory, black box, white box, and acceptance testing...getting usable feedback...tracking issues...choosing and creating tools...testing "Docs & Mocks," interfaces, classes, modules, libraries, binaries, services, and infrastructure...reviewing code and refactoring...using test hooks, presubmit scripts, queues, continuous builds, and more. With these techniques, you can transform testing from a bottleneck into an accelerator-and make your whole organization more productive!
Programming Finite Elements in Java (TM) teaches the reader how to programme the algorithms of the finite element method (FEM) in Java (TM). The compact, simple code helps the student to read the algorithms, to understand them and thus to be able to refine them. All of the main aspects of finite element techniques are considered: finite element solution; generation of finite element meshes; and visualization of finite element models and results with Java 3D (TM). The step-by-step presentation includes algorithm programming and code explanation at each point. Problems and exercises are provided for each chapter, with Java (TM) source code and problem data sets available from http://extras.springer.com/2010/978-1-84882-971-8.
The original edition of this book was celebrated for its coverage of the central concepts of practical optimization techniques. This updated edition expands and illuminates the connection between the purely analytical character of an optimization problem, expressed by properties of the necessary conditions, and the behavior of algorithms used to solve a problem. Incorporating modern theoretical insights, this classic text is even more useful.
Correcting the Great Mistake People often mistake one thing for another. That's human nature. However, one would expect the leaders in a particular ?eld of endeavour to have superior ab- ities to discriminate among the developments within that ?eld. That is why it is so perplexing that the technology elite - supposedly savvy folk such as software developers, marketers and businessmen - have continually mistaken Web-based graphics for something it is not. The ?rst great graphics technology for the Web, VRML, has been mistaken for something else since its inception. Viewed variously as a game system, a format for architectural walkthroughs, a platform for multi-user chat and an augmentation of reality, VRML may qualify as the least understood invention in the history of inf- mation technology. Perhaps it is so because when VRML was originally introduced it was touted as a tool for putting the shopping malls of the world online, at once prosaic and horrifyingly mundane to those of us who were developing it. Perhaps those ?rst two initials,"VR,"created expectations of sprawling, photorealistic f- tasy landscapes for exploration and play across the Web. Or perhaps the magnitude of the invention was simply too great to be understood at the time by the many, ironically even by those spending the money to underwrite its development. Regardless of the reasons, VRML suffered in the mainstream as it was twisted to meet unintended ends and stretched far beyond its limitations.
Defining a new development life-cycle methodology, together with a set of associated techniques and tools to develop highly critical systems using formal techniques, this book adopts a rigorous safety assessment approach explored via several layers (from requirements analysis to automatic source code generation). This is assessed and evaluated via a standard case study: the cardiac pacemaker. Additionally a formalisation of an Electrocardiogram (ECG) is used to identify anomalies in order to improve existing medical protocols. This allows the key issue - that formal methods are not currently integrated into established critical systems development processes - to be discussed in a highly effective and informative way. "Using Event-B for Critical Device Software Systems" serves as a valuable resource for researchers and students of formal methods. The assessment of critical systems development is applicable to all industries, but engineers and physicians from the health domain will find the cardiac pacemaker case study of particular value.
Recently, rapid and fundamental advances in computing technologies have been driving the role and scope of software systems to a new level. A number of new types of software systems are emerging, among which service based systems, cloud computing, pervasive computing, and Internet of Things are eminent examples. These systems have imposed new challenges on their design, development, testing and maintenance, and are demanding for new appropriate approaches and tools. Advanced Design Approaches to Emerging Software Systems: Principles, Methodology and Tools provides relevant theoretical frameworks and the latest empirical research findings in the area. This book will help to clarify the present chaotic and confusing literature of the current state of the art and knowledge in the areas of the design and engineering of the many emerging software systems. This book also facilitates the exchange and evolution of software engineering advances among multiple disciplines, research, industry, and user communities.
"The Grid is an emerging infrastructure that will fundamentally
change the way we think about-and use-computing. The word Grid is
used by analogy with the electric power grid, which provides
pervasive access to electricity and has had a dramatic impact on
human capabilities and society. Many believe that by allowing all
components of our information technology
infrastructure-computational capabilities, databases, sensors, and
people-to be shared flexibly as true collaborative tools the Grid
will have a similar transforming effect, allowing new classes of
applications to emerge."
Papers in this unique volume were developed from the 2006 conference hosted by IBM, Service Science, Management, and Engineering (SSME) Education for the 21st Century. The book incorporates a variety of perspectives, informed by an international background in SSME experience and education, including management, business, social science, computer science and engineering. Readers will derive an understanding of education needs and program offerings in SSME.
DSSSL (Document Style Semantics and Specification Language) is an ISO standard (ISO/IEC 10179: 1996) published in the year 1996. DSSSL is a standard of the SGML family (Standard Generalized Markup Language, ISO 8879:1986), whose aim is to establish a processing model for SGML documents. For a good understanding of the SGML standard, many books exist including Author's guide[BryanI988] and The SGML handbook[GoldfarbI990]. A DSSSL document is an SGML document, written with the same rules that guide any SGML document. The structure of a DSSSL document is explained in Chapter 2. DSSSL is based, in part, on scheme, a standard functional programming language. The DSSSL subset of scheme along with the procedures supported by DSSSL are explained in Chapter 3. The DSSSL standard starts with the supposition of a pre-existing SGML document, and offers a series of processes that can be performed on it: * Groves The first process that is performed on an SGML document in DSSSL is always the analysis of the document and the creation of a grove. The DSSSL standard shares many common characteristics with another standard of the SGML family, HyTime (ISO/IEC 10744). These standards were developed in parallel, and their developers designed a common data model, the grove, that would support the processing needs of each standard.
Dr. Tom and Debra Shinder have become synonymous with Microsoft's
flagship firewall product ISA Server, as a result of Tom's
prominent role as a member of the beta development team, and Tom
and Deb's featured placement on both Microsoft's ISA Server Web
site and ISAserver.org. Tom and Deb's book on the first release of
the product "Configuring ISA Server 2000" dominated the ISA Server
2000 book market having sold over 40,000 copies worldwide, and the
ISA Server community is eagerly awaiting Tom and Deb's book on ISA
Server 2004, which is the dramatically upgraded new release from
Microsoft.
This book develops a crowdsourced sensor-cloud service composition framework taking into account spatio-temporal aspects. This book also unfolds new horizons to service-oriented computing towards the direction of crowdsourced sensor data based applications, in the broader context of Internet of Things (IoT). It is a massive challenge for the IoT research field how to effectively and efficiently capture, manage and deliver sensed data as user-desired services. The outcome of this research will contribute to solving this very important question, by designing a novel service framework and a set of unique service selection and composition frameworks. Delivering a novel service framework to manage crowdsourced sensor data provides high-level abstraction (i.e., sensor-cloud service) to model crowdsourced sensor data from functional and non-functional perspectives, seamlessly turning the raw data into "ready to go" services. A creative indexing model is developed to capture and manage the spatio-temporal dynamism of crowdsourced service providers. Delivering novel frameworks to compose crowdsourced sensor-cloud services is vital. These frameworks focuses on spatio-temporal composition of crowdsourced sensor-cloud services, which is a new territory for existing service oriented computing research. A creative failure-proof model is also designed to prevent composition failure caused by fluctuating QoS. Delivering an incentive model to drive the coverage of crowdsourced service providers is also vital. A new spatio-temporal incentive model targets changing coverage of the crowdsourced providers to achieve demanded coverage of crowdsourced sensor-cloud services within a region. The outcome of this research is expected to potentially create a sensor services crowdsourcing market and new commercial opportunities focusing on crowdsourced data based applications. The crowdsourced community based approach adds significant value to journey planning and map services thus creating a competitive edge for a technologically-minded companies incentivizing new start-ups, thus enabling higher market innovation. This book primarily targets researchers and practitioners, who conduct research work in service oriented computing, Internet of Things (IoT), smart city and spatio-temporal travel planning, as well as advanced-level students studying this field. Small and Medium Entrepreneurs, who invest in crowdsourced IoT services and journey planning infrastructures, will also want to purchase this book.
The Stellent Content Server has a flexible and highly extensible service-oriented architecture. It is relatively easy to develop custom applications with it, or make modifications to the existing interface. However, there has never been a good manual on how to make such applications. Thata (TM)s where this book comes in: to help administrators and web developers create custom applications with the Content Server. It contains step-by-step instructions on how to create many different types of applications, including simple business forms, new web interfaces for existing services, new skins, and new functionality with custom Java code. It also contains detailed reference material for the inner workings of the Content Server. Youa (TM)ll need this vital resource when creating custom Stellent applications, and youa (TM)ll find this information only in this complete guide to Stellent Content Server.
Agile methods have taken software development by storm - but a growing controversy over their true effectiveness now pits proponents against detractors in a heated battle of claims. Here at last is a completely impartial guide that gives project managers both expert objective analysis of Agile software development methods and much-needed tools for evaluating the suitability of Agile methods for their organization. The book reviews the Agile Manifesto and the 12 Agile Principles, and discusses in detail each practice of the six most widely recognized Agile methods. Software developers get clear and unbiased insight into adoption implications, the benefits that may accrue, and the potential pitfalls they face. At its core, this unique assessment resource helps developers systematically address key Agile adoption issues involving organizational culture, customers, staff, tools and processes, and project considerations. A comprehensive start-to-finish worksheet makes it easy for professionals to document observations, draw justifiable conclusions, and market their recommendations to the company. The guide also features expert guidelines for customizing an Agile method, training staff, testing the new method, and evaluating its performance - should readers go the Agile route. Practical ready-to-use tools, including equations and templates in Excel format, are provided in a handy workbook which can be accessed through the author's website.
This book is a compilation of a selected subset of research articles presented at the Eighth INFORMS Computing Society Conference, held in Chandler, Arizona, from January 8 to 10, 2003. The articles in this book represent the diversity and depth of the interface between ORiMS (operations research and the management sciences) and CS/AI (computer science and artificial intelligence ). This volume starts with two papers that represent the reflective and integrative thinking that is critical to any scientific discipline. These two articles present philosophical perspectives on computation, covering a variety of traditional and newer methods for modeling, solving, and explaining mathematical models. The next set includes articles that study machine learning and computational heuristics, and is followed by articles that address issues in performance testing of solution algorithms and heuristics. These two sets of papers demonstrate the richness of thought that takes place at the ORiMS and CSI AI interface. The final set of articles demonstrates the usefulness of these and other methods at the interface towards solving problems in the real world, covering e-commerce, workflow, electronic negotiation, music, parallel computation, and telecommunications. The articles in this collection represent the results of cross-fertilization between ORiMS and CSI AI, making possible advances that could have not been achieved in isolation. The continuing aim ofthe INFORMS Computing Society and this research conference is to invigorate and further develop this interface.
For the first time, this up-to-date text combines the main issues of the hardware description language VHDL-AMS aimed at model representation of mixed-signal circuits and systems, characterization methods and tools for the extraction of model parameters, and modelling methodologies for accurate high-level behavioural models.
Managing risk is essential for every organization. However, significant opportunities may be lost by concentrating on the negative aspects of risk without bearing in mind the positive attributes. The objective of Project Risk Management: Managing Software Development Risk is to provide a distinct approach to a broad range of risks and rewards associated with the design, development, implementation and deployment of software systems. The traditional perspective of software development risk is to view risk as a negative characteristic associated with the impact of potential threats. The perspective of this book is to explore a more discerning view of software development risks, including the positive aspects of risk associated with potential beneficial opportunities. A balanced approach requires that software project managers approach negative risks with a view to reduce the likelihood and impact on a software project, and approach positive risks with a view to increase the likelihood of exploiting opportunities. Project Risk Management: Managing Software Development Risk explores software development risk both from a technological and business perspective. Issues regarding strategies for software development are discussed and topics including risks related to technical performance, outsourcing, cybersecurity, scheduling, quality, costs, opportunities and competition are presented. Bringing together concepts across the broad spectrum of software engineering with a project management perspective, this volume represents both a professional and scholarly perspective on the topic.
The information infrastructure---comprising computers, embedded devices, networks and software systems---is vital to day-to-day operations in every sector: information and telecommunications, banking and finance, energy, chemicals and hazardous materials, agriculture, food, water, public health, emergency services, transportation, postal and shipping, government and defense. Global business and industry, governments, indeed society itself, cannot function effectively if major components of the critical information infrastructure are degraded, disabled or destroyed. Critical Infrastructure Protection V describes original research results and innovative applications in the interdisciplinary field of critical infrastructure protection. Also, it highlights the importance of weaving science, technology and policy in crafting sophisticated, yet practical, solutions that will help secure information, computer and network assets in the various critical infrastructure sectors. Areas of coverage include: Themes and Issues, Control Systems Security, Infrastructure Security, and Infrastructure Modeling and Simulation. This book is the 5th volume in the annual series produced by the International Federation for Information Processing (IFIP) Working Group 11.10 on Critical Infrastructure Protection, an international community of scientists, engineers, practitioners and policy makers dedicated to advancing research, development and implementation efforts focused on infrastructure protection. The book contains a selection of 14 edited papers from the 5th Annual IFIP WG 11.10 International Conference on Critical Infrastructure Protection, held at Dartmouth College, Hanover, New Hampshire, USA in the spring of 2011. Critical Infrastructure Protection V is an important resource for researchers, faculty members and graduate students, as well as for policy makers, practitioners and other individuals with interests in homeland security. Jonathan Butts is an Assistant Professor of Computer Science at the Air Force Institute of Technology, Wright-Patterson Air Force Base, Ohio, USA. Sujeet Shenoi is the F.P. Walter Professor of Computer Science at the University of Tulsa, Tulsa, Oklahoma, USA.
The multicore revolution has reached the deployment stage in
embedded systems ranging from small ultramobile devices to large
telecommunication servers. The transition from single to multicore
processors, motivated by the need to increase performance while
conserving power, has placed great responsibility on the shoulders
of software engineers. In this new embedded multicore era, the
toughest task is the development of code to support more
sophisticated systems. This book provides embedded engineers with
solid grounding in the skills required to develop software
targeting multicore processors. Within the text, the author
undertakes an in-depth exploration of performance analysis, and a
close-up look at the tools of the trade. Both general multicore
design principles and processor-specific optimization techniques
are revealed. Detailed coverage of critical issues for multicore
employment within embedded systems is provided, including the
Threading Development Cycle, with discussions of analysis, design,
development, debugging, and performance tuning of threaded
applications. Software development techniques engendering optimal
mobility and energy efficiency are highlighted through multiple
case studies, which provide practical how-to advice on implementing
the latest multicore processors. Finally, future trends are
discussed, including terascale, speculative multithreading,
transactional memory, interconnects, and the software-specific
implications of these looming architectural developments.
The Handbook of Service Description introduces an in-depth overview of service description efforts. The book also highlights the recent Unified Service Description Language (USDL) in detail and discusses its methods. The Handbook of Service Description is the normative scientific reference for the upcoming standardization of the Unified Service Description Language (USDL). Complete documentation is included. The Handbook of Service Description is designed for those working in the service science industry as a reference book. Advanced-level students focused on computer science, engineering and business will also find this book a valuable asset. |
You may like...
Java How to Program, Late Objects…
Paul Deitel, Harvey Deitel
Paperback
News Search, Blogs and Feeds - A Toolkit
Lars Vage, Lars Iselid
Paperback
R1,332
Discovery Miles 13 320
Introducing Delphi Programming - Theory…
John Barrow, Linda Miller, …
Paperback
(1)R785 Discovery Miles 7 850
Hardware Accelerator Systems for…
Shiho Kim, Ganesh Chandra Deka
Hardcover
R3,950
Discovery Miles 39 500
|