Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > Computer programming > General
Optimization is a field important in its own right but is also integral to numerous applied sciences, including operations research, management science, economics, finance and all branches of mathematics-oriented engineering. Constrained optimization models are one of the most widely used mathematical models in operations research and management science. This book gives a modern and well-balanced presentation of the subject, focusing on theory but also including algorithims and examples from various real-world applications. The text is easy to read and accessible to anyone with a knowledge of multi-dimensional calculus, linear algebra and basic numerical methods. Detailed examples and counter-examples are provided--as are exercises, solutions and helpful hints, and Matlab/Maple supplements. The intended readership is advanced undergraduates, graduates, and professionals in any of the applied fields.
This text on geometry is devoted to various central geometrical topics including: graphs of functions, transformations, (non-)Euclidean geometries, curves and surfaces as well as their applications in a variety of disciplines. This book presents elementary methods for analytical modeling and demonstrates the potential for symbolic computational tools to support the development of analytical solutions. The author systematically examines several powerful tools of MATLAB (R) including 2D and 3D animation of geometric images with shadows and colors and transformations using matrices. With over 150 stimulating exercises and problems, this text integrates traditional differential and non-Euclidean geometries with more current computer systems in a practical and user-friendly format. This text is an excellent classroom resource or self-study reference for undergraduate students in a variety of disciplines.
A Modular Calculus for the Average Cost of Data Structuring introduces MOQA, a new domain-specific programming language which guarantees the average-case time analysis of its programs to be modular.Time in this context refers to a broad notion of cost, which can be used to estimate the actual running time, but also other quantitative information such as power consumption, while modularity means that the average time of a program can be easily computed from the times of its constituents--something that no programming language of this scope has been able to guarantee so far. MOQA principles can be incorporated in any standard programming language. MOQA supports tracking of data and their distributions throughout computations, based on the notion of random bag preservation. This allows a unified approach to average-case time analysis, and resolves fundamental bottleneck problems in the area. The main techniques are illustrated in an accompanying Flash tutorial, where the visual nature of this method can provide new teaching ideas for algorithms courses. This volume, with forewords by Greg Bollella and Dana Scott, presents novel programs based on the new advances in this area, including the first randomness-preserving version of Heapsort. Programs are provided, along with derivations of their average-case time, to illustrate the radically different approach to average-case timing. The automated static timing tool applies the Modular Calculus to extract the average-case running time of programs directly from their MOQA code. A Modular Calculus for the Average Cost of Data Structuring is designed for a professional audience composed of researchers and practitioners in industry, with an interest in algorithmic analysis and also static timing and power analysis--areas of growing importance. It is also suitable as an advanced-level text or reference book for students in computer science, electrical engineering and mathematics. Michel Schellekens obtained his PhD from Carnegie Mellon University, following which he worked as a Marie Curie Fellow at Imperial College London. Currently he is an Associate Professor at the Department of Computer Science in University College Cork - National University of Ireland, Cork, where he leads the Centre for Efficiency-Oriented Languages (CEOL) as a Science Foundation Ireland Principal Investigator.
This book presents works detailing the application of processing and visualization techniques for analyzing the Earth's subsurface. The topic of the book is interactive data processing and interactive 3D visualization techniques used on subsurface data. Interactive processing of data together with interactive visualization is a powerful combination which has in the recent years become possible due to hardware and algorithm advances in. The combination enables the user to perform interactive exploration and filtering of datasets while simultaneously visualizing the results so that insights can be made immediately. This makes it possible to quickly form hypotheses and draw conclusions. Case studies from the geosciences are not as often presented in the scientific visualization and computer graphics community as e.g., studies on medical, biological or chemical data. This book will give researchers in the field of visualization and computer graphics valuable insight into the open visualization challenges in the geosciences, and how certain problems are currently solved using domain specific processing and visualization techniques. Conversely, readers from the geosciences will gain valuable insight into relevant visualization and interactive processing techniques. Subsurface data has interesting characteristics such as its solid nature, large range of scales and high degree of uncertainty, which makes it challenging to visualize with standard methods. It is also noteworthy that parallel fields of research have taken place in geosciences and in computer graphics, with different terminology when it comes to representing geometry, describing terrains, interpolating data and (example-based) synthesis of data. The domains covered in this book are geology, digital terrains, seismic data, reservoir visualization and CO2 storage. The technologies covered are 3D visualization, visualization of large datasets, 3D modelling, machine learning, virtual reality, seismic interpretation and multidisciplinary collaboration. People within any of these domains and technologies are potential readers of the book.
This book introduces the reader with little or no previous computer-programming experience to the Python programming language of interest for a physicist or a natural-sciences student. The book starts with basic interactive Python in order to acquire an introductory familiarity with the language, than tackle Python scripts (programs) of increasing complexity, that the reader is invited to run on her/his computer. All program listings are discussed in detail, and the reader is invited to experiment on what happens if some code lines are modified. The reader is introduced to Matplotlib graphics for the generation of figures representing data and function plots and, for instance, field lines. Animated function plots are also considered. A chapter is dedicated to the numerical solution of algebraic and transcendental equations, the basic mathematical principles are discussed and the available Python tools for the solution are presented. A further chapter is dedicated to the numerical solution of ordinary differential equations. This is of vital importance for the physicist, since differential equations are at the base of both classical physics (Newton's equations) and quantum mechanics (Schroedinger's equation). The shooting method for the numerical solution of ordinary differential equations with boundary conditions at two boundaries is also presented. Python programs for the solution of two quantum-mechanics problems are discussed as examples. Two chapters are dedicated to Tkinter graphics, which gives the user more freedom than Matplotlib, and to Tkinter animation. Programs displaying the animation of physical problems involving the solution of ordinary differential equations (for which in most cases there is no algebraic solution) in real time are presented and discussed. Finally, 3D animation is presented with Vpython.
This work, a tribute to renowned researcher Robert Paige, is a collection of revised papers published in his honor in the Higher-Order and Symbolic Computation Journal in 2003 and 2005. Among them there are two key papers: a retrospective view of his research lines, and a proposal for future studies in the area of the automatic program derivation. The book also includes some papers by members of the IFIP Working Group 2.1 of which Bob was an active member.
Julia Kay's Portrait Party is an international collaborative project in which artists all over the world make portraits of each other and share them online. After years of exchanging portraits, tips and techniques within the group, in Portrait Revolution these artists are now sharing their art, their words, and their inspiration with everyone who is interested in or would like to get started with portraiture. Here you can find information on using different media, how to handle difficult portrait issues, and more. Portrait Revolution showcases 450 portraits by 200 artists, in a wide variety of media from oil painting to iPad art, watercolour to ballpoint, linocut to mosaic. There are a range of styles from realistic to abstract and interpretations by multiple artists of the same subject.
Broadband Satellite Communication Systems and the Challenges of Mobility is an essential reference for both academic and professional researchers in the field of telecommunications, computer networking and wireless networks. Recently the request of multimedia services has been rapidly increasing and satellite networks appear to be attractive for a fast service deployment and for extending the typical service area of terrestrial systems. In comparison with traditional wide area networks, a characteristic of satellite communication systems is their ability in broadcasting and multicasting multimedia information flows anywhere over the satellite coverage. The papers presented in this volume highlight key areas such as Satellite Network Architectures, Services and Applications; Mobile Satellite Systems and Services; and Hybrid Satellite and Terrestrial Networks. Mobility will inevitably be one of the main characteristics of future networks, terminals and applications and, thus, extending and integrating fixed network protocols and services to mobile systems represents one of the main issues of present networking. The secondary focus of this volume is on challenges of mobility, that is, on technologies, protocols and services for the support of seamless and nomadic user access to new classes of applications in person-to-person, device-to-device and device-to-person environments. The book comprises recent results of research and development in the following areas; Seamless mobility; Mobile ad hoc and sensor networks; Analysis, simulation and measurements of mobile and wireless systems; Integration and inter-working of wired and wireless networks; QoS in mobile and wireless networks; Future trends and issues concerning mobility. This state -of-the-art volume contains a collection of papers from two of the workshops of the 18th IFIP World Computer Congress, held August 22-27, 2004, in Toulouse, France: the Workshop on Broadband Satellite Communication Systems, and the Workshop on the Challenges of Mobility.
The design of formal calculi in which fundamental concepts underlying interactive systems can be described and studied has been a central theme of theoretical computer science in recent decades, while membrane computing, a rule-based formalism inspired by biological cells, is a more recent field that belongs to the general area of natural computing. This is the first book to establish a link between these two research directions while treating mobility as the central topic. In the first chapter the authors offer a formal description of mobility in process calculi, noting the entities that move: links ( -calculus), ambients (ambient calculi) and branes (brane calculi). In the second chapter they study mobility in the framework of natural computing. The authors define several systems of mobile membranes in which the movement inside a spatial structure is provided by rules inspired by endocytosis and exocytosis. They study their computational power in comparison with the classical notion of Turing computability and their efficiency in algorithmically solving hard problems in polynomial time. The final chapter deals with encodings, establishing links between process calculi and membrane computing so that researchers can share techniques between these fields. The book is suitable for computer scientists working in concurrency and in biologically inspired formalisms, and also for mathematically inclined scientists interested in formalizing moving agents and biological phenomena. The text is supported with examples and exercises, so it can also be used for courses on these topics.
This volume provides an overview of current work in software engineering techniques that can enhance the quality of software. The chapters of this volume, organized by key topic area, create an agenda for the IFIP Working Conference on Software Engineering Techniques, SET 2006. The seven sections of the volume address the following areas: software architectures, modeling, project management, software quality, analysis and verification methods, data management, and software maintenance.
Austin Sincock provides step-by-step real-world examples for developing Enterprise Java applications on SAP. His is the first title that uses open-source software to help developers learn and use Java for SAP. Bridges the gap between SAP's language ABAP and object-oriented Java Provides a complete look at SAP's Java connector, JCo Demonstrates graphical application development for SAP using Java's Swing libraries Shows how to deploy and build Java applications on the Tomcat Java application server Teaches how to deploy the Java-based HypersonicSQL database, use SQL to populate the database, and tie the external database into an SAP system with Java
This book develops a crowdsourced sensor-cloud service composition framework taking into account spatio-temporal aspects. This book also unfolds new horizons to service-oriented computing towards the direction of crowdsourced sensor data based applications, in the broader context of Internet of Things (IoT). It is a massive challenge for the IoT research field how to effectively and efficiently capture, manage and deliver sensed data as user-desired services. The outcome of this research will contribute to solving this very important question, by designing a novel service framework and a set of unique service selection and composition frameworks. Delivering a novel service framework to manage crowdsourced sensor data provides high-level abstraction (i.e., sensor-cloud service) to model crowdsourced sensor data from functional and non-functional perspectives, seamlessly turning the raw data into "ready to go" services. A creative indexing model is developed to capture and manage the spatio-temporal dynamism of crowdsourced service providers. Delivering novel frameworks to compose crowdsourced sensor-cloud services is vital. These frameworks focuses on spatio-temporal composition of crowdsourced sensor-cloud services, which is a new territory for existing service oriented computing research. A creative failure-proof model is also designed to prevent composition failure caused by fluctuating QoS. Delivering an incentive model to drive the coverage of crowdsourced service providers is also vital. A new spatio-temporal incentive model targets changing coverage of the crowdsourced providers to achieve demanded coverage of crowdsourced sensor-cloud services within a region. The outcome of this research is expected to potentially create a sensor services crowdsourcing market and new commercial opportunities focusing on crowdsourced data based applications. The crowdsourced community based approach adds significant value to journey planning and map services thus creating a competitive edge for a technologically-minded companies incentivizing new start-ups, thus enabling higher market innovation. This book primarily targets researchers and practitioners, who conduct research work in service oriented computing, Internet of Things (IoT), smart city and spatio-temporal travel planning, as well as advanced-level students studying this field. Small and Medium Entrepreneurs, who invest in crowdsourced IoT services and journey planning infrastructures, will also want to purchase this book.
The book presents contributions on statistical models and methods applied, for both data science and SDGs, in one place. Measuring and controlling data of SDGs, data driven measurement of progress needs to be distributed to stakeholders. In this situation, the techniques used in data science, specially, in the big data analytics, play an important role rather than the traditional data gathering and manipulation techniques. This book fills this space through its twenty contributions. The contributions have been selected from those presented during the 7th International Conference on Data Science and Sustainable Development Goals organized by the Department of Statistics, University of Rajshahi, Bangladesh; and cover topics mainly on SDGs, bioinformatics, public health, medical informatics, environmental statistics, data science and machine learning. The contents of the volume would be useful to policymakers, researchers, government entities, civil society, and nonprofit organizations for monitoring and accelerating the progress of SDGs.
Recently, rapid and fundamental advances in computing technologies have been driving the role and scope of software systems to a new level. A number of new types of software systems are emerging, among which service based systems, cloud computing, pervasive computing, and Internet of Things are eminent examples. These systems have imposed new challenges on their design, development, testing and maintenance, and are demanding for new appropriate approaches and tools. Advanced Design Approaches to Emerging Software Systems: Principles, Methodology and Tools provides relevant theoretical frameworks and the latest empirical research findings in the area. This book will help to clarify the present chaotic and confusing literature of the current state of the art and knowledge in the areas of the design and engineering of the many emerging software systems. This book also facilitates the exchange and evolution of software engineering advances among multiple disciplines, research, industry, and user communities.
."NET 2.0 for Delphi Programmers" explores .NET from a Delphi programmers viewpoint, and it is ideal for Delphi programmers moving to .NET. It presents the core concepts of the .NET world in terms you are familiar with. This book will help you with Delphi for .NET as well as C#. Apress publishes migration books for both VB 6 and C++ programmers moving to .NET. Consider this the Delphi installment of Apress migration books! There is ample coverage of C# as well as Delphi for .NET inside this newest addition.
Correcting the Great Mistake People often mistake one thing for another. That's human nature. However, one would expect the leaders in a particular ?eld of endeavour to have superior ab- ities to discriminate among the developments within that ?eld. That is why it is so perplexing that the technology elite - supposedly savvy folk such as software developers, marketers and businessmen - have continually mistaken Web-based graphics for something it is not. The ?rst great graphics technology for the Web, VRML, has been mistaken for something else since its inception. Viewed variously as a game system, a format for architectural walkthroughs, a platform for multi-user chat and an augmentation of reality, VRML may qualify as the least understood invention in the history of inf- mation technology. Perhaps it is so because when VRML was originally introduced it was touted as a tool for putting the shopping malls of the world online, at once prosaic and horrifyingly mundane to those of us who were developing it. Perhaps those ?rst two initials,"VR,"created expectations of sprawling, photorealistic f- tasy landscapes for exploration and play across the Web. Or perhaps the magnitude of the invention was simply too great to be understood at the time by the many, ironically even by those spending the money to underwrite its development. Regardless of the reasons, VRML suffered in the mainstream as it was twisted to meet unintended ends and stretched far beyond its limitations.
The past decades have seen significant improvements in 3D imaging where the related techniques and technologies have advanced to a mature state. These exciting developments have sparked increasing interest in the challenges and opportunities afforded by 3D sensing. As a consequence, the emerging area of safety and security related imaging incorporates these important new technologies beyond the limitations of 2D image processing.This book presents the thoroughly revised versions of lectures given by leading researchers during the Workshop on Advanced 3D Imaging for Safety and Security in conjunction with the International Conference on Computer Vision and Pattern Recognition CVPR 2005, held in San Diego, CA, USA in June 2005.It covers the current state of the art in 3D imaging for safety and security.
The Stellent Content Server has a flexible and highly extensible service-oriented architecture. It is relatively easy to develop custom applications with it, or make modifications to the existing interface. However, there has never been a good manual on how to make such applications. Thata (TM)s where this book comes in: to help administrators and web developers create custom applications with the Content Server. It contains step-by-step instructions on how to create many different types of applications, including simple business forms, new web interfaces for existing services, new skins, and new functionality with custom Java code. It also contains detailed reference material for the inner workings of the Content Server. Youa (TM)ll need this vital resource when creating custom Stellent applications, and youa (TM)ll find this information only in this complete guide to Stellent Content Server.
Maintaining SQL Server 2005 high availability in a global information environment is the database administrators greatest practical challenge. Availability is as much about people and processes as it is about technology. Microsoft SQL Server 2005 High Availability covers the technology, people, processes, and real-world best practices for planning, deploying, administering, and maintaining highly available SQL Server 2005 instances. With years of experience as a database consultant and production DBA, Allan Hirt provides in-depth, detailed advice on what it takes to ensure SQL Server high availability for businesses of any size. This is not an academic text; its not based on lab experiments, but on real-world experience. This book is a dramatic update and revision of the authors previous best-seller on SQL Server 2000. It gives sound guidance to DBAs and system administrators on how to really get the job done.
The original edition of this book was celebrated for its coverage of the central concepts of practical optimization techniques. This updated edition expands and illuminates the connection between the purely analytical character of an optimization problem, expressed by properties of the necessary conditions, and the behavior of algorithms used to solve a problem. Incorporating modern theoretical insights, this classic text is even more useful.
The growing commercial market of Microwave/ Millimeter wave industry over the past decade has led to the explosion of interests and opportunities for the design and development of microwave components.The design of most microwave components requires the use of commercially available electromagnetic (EM) simulation tools for their analysis. In the design process, the simulations are carried out by varying the design parameters until the desired response is obtained. The optimization of design parameters by manual searching is a cumbersome and time consuming process. Soft computing methods such as Genetic Algorithm (GA), Artificial Neural Network (ANN) and Fuzzy Logic (FL) have been widely used by EM researchers for microwave design since last decade. The aim of these methods is to tolerate imprecision, uncertainty, and approximation to achieve robust and low cost solution in a small time frame. Modeling and optimization are essential parts and powerful tools for the microwave/millimeter wave design. This book deals with the development and use of soft computing methods for tackling challenging design problems in the microwave/millimeter wave domain. The aim in the development of these methods is to obtain the design in small time frame while improving the accuracy of the design for a wide range of applications. To achieve this goal, a few diverse design problems of microwave field, representing varied challenges in the design, such as different microstrip antennas, microwave filters, a microstrip-via and also some critical high power components such as nonlinear tapers and RF-windows are considered as case-study design problems. Different design methodologies are developed for these applications. The presents soft computing methods, their review for microwave/millimeter wave design problems and specific case-study problems to infuse better insight and understanding of the subject.
DSSSL (Document Style Semantics and Specification Language) is an ISO standard (ISO/IEC 10179: 1996) published in the year 1996. DSSSL is a standard of the SGML family (Standard Generalized Markup Language, ISO 8879:1986), whose aim is to establish a processing model for SGML documents. For a good understanding of the SGML standard, many books exist including Author's guide[BryanI988] and The SGML handbook[GoldfarbI990]. A DSSSL document is an SGML document, written with the same rules that guide any SGML document. The structure of a DSSSL document is explained in Chapter 2. DSSSL is based, in part, on scheme, a standard functional programming language. The DSSSL subset of scheme along with the procedures supported by DSSSL are explained in Chapter 3. The DSSSL standard starts with the supposition of a pre-existing SGML document, and offers a series of processes that can be performed on it: * Groves The first process that is performed on an SGML document in DSSSL is always the analysis of the document and the creation of a grove. The DSSSL standard shares many common characteristics with another standard of the SGML family, HyTime (ISO/IEC 10744). These standards were developed in parallel, and their developers designed a common data model, the grove, that would support the processing needs of each standard.
This book is a compilation of a selected subset of research articles presented at the Eighth INFORMS Computing Society Conference, held in Chandler, Arizona, from January 8 to 10, 2003. The articles in this book represent the diversity and depth of the interface between ORiMS (operations research and the management sciences) and CS/AI (computer science and artificial intelligence ). This volume starts with two papers that represent the reflective and integrative thinking that is critical to any scientific discipline. These two articles present philosophical perspectives on computation, covering a variety of traditional and newer methods for modeling, solving, and explaining mathematical models. The next set includes articles that study machine learning and computational heuristics, and is followed by articles that address issues in performance testing of solution algorithms and heuristics. These two sets of papers demonstrate the richness of thought that takes place at the ORiMS and CSI AI interface. The final set of articles demonstrates the usefulness of these and other methods at the interface towards solving problems in the real world, covering e-commerce, workflow, electronic negotiation, music, parallel computation, and telecommunications. The articles in this collection represent the results of cross-fertilization between ORiMS and CSI AI, making possible advances that could have not been achieved in isolation. The continuing aim ofthe INFORMS Computing Society and this research conference is to invigorate and further develop this interface.
XML is bringing together some fairly disparate groups into a new
cultural clash: document developers trying to understand what a
transaction is, database analysts getting upset because the
relational model doesn't fit anymore, and web designers having to
deal with schemata and rule based transformations. The key to
rising above the confusion is to understand the different semantic
structures that lie beneath the standards of XML, and how to model
the semantics to achieve the goals of the organization.
This handbook offers comprehensive coverage of recent advancements in Big Data technologies and related paradigms. Chapters are authored by international leading experts in the field, and have been reviewed and revised for maximum reader value. The volume consists of twenty-five chapters organized into four main parts. Part one covers the fundamental concepts of Big Data technologies including data curation mechanisms, data models, storage models, programming models and programming platforms. It also dives into the details of implementing Big SQL query engines and big stream processing systems. Part Two focuses on the semantic aspects of Big Data management including data integration and exploratory ad hoc analysis in addition to structured querying and pattern matching techniques. Part Three presents a comprehensive overview of large scale graph processing. It covers the most recent research in large scale graph processing platforms, introducing several scalable graph querying and mining mechanisms in domains such as social networks. Part Four details novel applications that have been made possible by the rapid emergence of Big Data technologies such as Internet-of-Things (IOT), Cognitive Computing and SCADA Systems. All parts of the book discuss open research problems, including potential opportunities, that have arisen from the rapid progress of Big Data technologies and the associated increasing requirements of application domains. Designed for researchers, IT professionals and graduate students, this book is a timely contribution to the growing Big Data field. Big Data has been recognized as one of leading emerging technologies that will have a major contribution and impact on the various fields of science and varies aspect of the human society over the coming decades. Therefore, the content in this book will be an essential tool to help readers understand the development and future of the field. |
You may like...
SAS Graphics for Clinical Trials by…
Kriss Harris, Richann Watson
Hardcover
R1,578
Discovery Miles 15 780
Intro to Python for Computer Science and…
Paul Deitel
Paperback
An Introduction to XML and Web…
Anders Moller, Michael Schwartzbach
Paperback
R2,413
Discovery Miles 24 130
|