![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Computer programming
This book introduces all the relevant information required to understand and put Model Driven Architecture (MDA) into industrial practice. It clearly explains which conceptual primitives should be present in a system specification, how to use UML to properly represent this subset of basic conceptual constructs, how to identify just those diagrams and modeling constructs that are actually required to create a meaningful conceptual schema, and how to accomplish the transformation process between the problem space and the solution space. The approach is fully supported by commercially available tools.
The purpose of the 11th International Conference on Software Engineering Research, Management and Applications (SERA 2013) held on August 7 - 9, 2012 in Prague, Czech Republic was to bring together scientists, engineers, computer users, and students to share their experiences and exchange new ideas and research results about all aspects (theory, applications and tools) of Software Engineering Research, Management and Applications, and to discuss the practical challenges encountered along the way and the solutions adopted to solve them. The conference organizers selected 17 outstanding papers from those papers accepted for presentation at the conference in order to publish them in this volume. The papers were chosen based on review scores submitted by members of the program committee, and further rigorous rounds of review.
This book addresses patient-specific modeling. It integrates computational modeling, experimental procedures, imagine clinical segmentation and mesh generation with the finite element method (FEM) to solve problems in computational biomedicine and bioengineering. Specific areas of interest include cardiovascular problems, ocular and muscular systems and soft tissue modeling. Patient-specific modeling has been the subject of serious research over the last seven years and interest in the area is continually growing and this area is expected to further develop in the near future.
Recent developments in computer science clearly show the need for a
better theoretical foundation for some central issues. Methods and
results from mathematical logic, in particular proof theory and
model theory, are of great help here and will be used much more in
future than previously. This book provides an excellent
introduction to the interplay of mathematical logic and computer
science. It contains extensively reworked versions of the lectures
given at the 1997 Marktoberdorf Summer School by leading
researchers in the field.
Robust Technology with Analysis of Interference in Signal Processing discusses for the first time the theoretical fundamentals and algorithms of analysis of noise as an information carrier. On their basis the robust technology of noisy signals processing is developed. This technology can be applied to solving the problems of control, identification, diagnostics, and pattern recognition in petrochemistry, energetics, geophysics, medicine, physics, aviation, and other sciences and industries. The text explores the emergent possibility of forecasting failures on various objects, in conjunction with the fact that failures follow the hidden microchanges revealed via interference estimates. This monograph is of interest to students, postgraduates, engineers, scientific associates and others who are concerned with the processing of measuring information on computers.
On August 1997 a conference titled "From Local to Global Optimiza- tion" was held at Storgarden in Rimfor.sa near the Linkoping Institute of Technology, Sweden. The conference gave us the opportunity to cel- ebrate Hoang Thy's achievements in Optimization during his 70 years of life. This book consists of a collection of research papers based on results presented during the conference and are dedicated to Professor Hoang Thy on the occasion of his 70th birthday. The papers cover a wide range of recent results in Mathematical Pro- gramming. The work of Hoang Thy, in particular in Global Optimiza- tion, has provided directions for new algorithmic developments in the field. We are indebted to the Kluwer Academic Publishers for inviting us to publish this volume, and the Center for Industrial Information Transfer (CENIIT) for financial support. We wish to thank the referees for their help and the authors for their papers. We also wish to join all contributors of this book in expressing birthday wishes and gratitude to Hoang Thy for his inspiration, support, and friendship to all of us. Athanasios Migdalas, Panos M. Pardalos, and Peter Varbrand November 1998 xv Hoang Tuy: An Appreciation Its a pleasure for me as colleague and friend to take this opportunity to celebrate Hoang 'I\lY'S numerous contributions to the field of mathemat- ical programming.
For years, companies have rewarded their most effective engineers by suggesting they move to a management position. But treating management as the default (or only) path for an engineer with leadership ability doesn't serve the industry well. The staff engineer path allows you to contribute at a high level, with more free time to drive big projects, determine tech strategy, and raise everyone's skills. With this in-depth book, author Tanya Reilly shows you ways to master strategic thinking, manage difficult projects, and set the standard for technical work. You'll learn how to be a leader without direct authority, how to plan ahead so that you're making the right technical decisions, and how to make everyone around you better, all while still leaving you time to grow as an expert in your domain. In three parts, you'll explore the three pillars of an engineer's job: Big picture thinking: learn how to take a broad, strategic view when thinking about your work Project execution: dive into tactics and explore the practicalities of making projects succeed Being a positive influence: determine the standards for what "good engineering" means in your organization
The methodology described in this book is the result of many years of research experience in the field of synthesizable VHDL design targeting FPGA based platforms. VHDL was first conceived as a documentation language for ASIC designs. Afterwards, the language was used for the behavioral simulation of ASICs, and also as a design input for synthesis tools. VHDL is a rich language, but just a small subset of it can be used to write synthesizable code, from which a physical circuit can be obtained. Usually VHDL books describe both, synthesis and simulation aspects of the language, but in this book the reader is conducted just through the features acceptable by synthesis tools. The book introduces the subjects in a gradual and concise way, providing just enough information for the reader to develop their synthesizable digital systems in VHDL. The examples in the book were planned targeting an FPGA platform widely used around the world.
'Visual Languages for Interactive Computing' presents problems and methodologies related to the syntax, semantics, and ambiguities of visual languages.
This book provides an extensive review of three interrelated issues: land fragmentation, land consolidation, and land reallocation, and it presents in detail the theoretical background, design, development and application of a prototype integrated planning and decision support system for land consolidation. The system integrates geographic information systems (GIS) and artificial intelligence techniques including expert systems (ES) and genetic algorithms (GAs) with multi-criteria decision methods (MCDM), both multi-attribute (MADM) and multi-objective (MODM). The system is based on four modules for measuring land fragmentation; automatically generating alternative land redistribution plans; evaluating those plans; and automatically designing the land partitioning plan. The presented research provides a new scientific framework for land-consolidation planning both in terms of theory and practice, by presenting new findings and by developing better tools and methods embedded in an integrated GIS environment. It also makes a valuable contribution to the fields of GIS and spatial planning, as it provides new methods and ideas that could be applied to improve the former for the benefit of the latter in the context of planning support systems. From the 1960s, ambitious research activities set out to observe regarding IT-support of the complex and time consuming redistribution processes within land consolidation without any practically relevant results, until now. This scientific work is likely to close that gap. This distinguished publication is highly recommended to land consolidation planning experts, researchers and academics alike. Prof. Dr.-Ing. Joachim Thomas, Munster/ Germany Prof. Michael Batty, University College London"
With the purpose of building upon standard web technologies, open linked data serves as a useful way to connect previously unrelated data and to publish structured data on the web. The application of these elements leads to the creation of data commons called semantic web. Cases on Open-Linked Data and Semantic Web Applications brings together new theories, research findings and case studies which cover the recent developments and approaches towards applied open linked data and semantic web in the context of information systems. By enhancing the understanding of open linked data in business, science and information technologies, this reference source aims to be useful for academics, researchers, and practitioners. With the purpose of building upon standard web technologies, open linked data serves as a useful way to connect previously unrelated data and to publish structured data on the web. The application of these elements leads to the creation of data commons called semantic web.
Advanced approaches to software engineering and design are capable of solving complex computational problems and achieving standards of performance that were unheard of only decades ago. Handbook of Research on Emerging Advancements and Technologies in Software Engineering presents a comprehensive investigation of the most recent discoveries in software engineering research and practice, with studies in software design, development, implementation, testing, analysis, and evolution. Software designers, architects, and technologists, as well as students and educators, will find this book to be a vital and in-depth examination of the latest notable developments within the software engineering community.
From a review of the Second Edition
In this book, the author considers separable programming and, in particular, one of its important cases - convex separable programming. Some general results are presented, techniques of approximating the separable problem by linear programming and dynamic programming are considered. Convex separable programs subject to inequality/ equality constraint(s) and bounds on variables are also studied and iterative algorithms of polynomial complexity are proposed. As an application, these algorithms are used in the implementation of stochastic quasigradient methods to some separable stochastic programs. Numerical approximation with respect to I1 and I4 norms, as a convex separable nonsmooth unconstrained minimization problem, is considered as well. Audience: Advanced undergraduate and graduate students, mathematical programming/ operations research specialists.
This monograph addresses advances in representation learning, a cutting-edge research area of machine learning. Representation learning refers to modern data transformation techniques that convert data of different modalities and complexity, including texts, graphs, and relations, into compact tabular representations, which effectively capture their semantic properties and relations. The monograph focuses on (i) propositionalization approaches, established in relational learning and inductive logic programming, and (ii) embedding approaches, which have gained popularity with recent advances in deep learning. The authors establish a unifying perspective on representation learning techniques developed in these various areas of modern data science, enabling the reader to understand the common underlying principles and to gain insight using selected examples and sample Python code. The monograph should be of interest to a wide audience, ranging from data scientists, machine learning researchers and students to developers, software engineers and industrial researchers interested in hands-on AI solutions.
Covering all the essential components of Unix/Linux, including process management, concurrent programming, timer and time service, file systems and network programming, this textbook emphasizes programming practice in the Unix/Linux environment. Systems Programming in Unix/Linux is intended as a textbook for systems programming courses in technically-oriented Computer Science/Engineering curricula that emphasize both theory and programming practice. The book contains many detailed working example programs with complete source code. It is also suitable for self-study by advanced programmers and computer enthusiasts. Systems programming is an indispensable part of Computer Science/Engineering education. After taking an introductory programming course, this book is meant to further knowledge by detailing how dynamic data structures are used in practice, using programming exercises and programming projects on such topics as C structures, pointers, link lists and trees. This book provides a wide range of knowledge about computer systemsoftware and advanced programming skills, allowing readers to interface with operatingsystem kernel, make efficient use of system resources and develop application software.It also prepares readers with the needed background to pursue advanced studies inComputer Science/Engineering, such as operating systems, embedded systems, databasesystems, data mining, artificial intelligence, computer networks, network security,distributed and parallel computing.
These proceedings contain the papers of IFIP/SEC 2010. It was a special honour and privilege to chair the Program Committee and prepare the proceedings for this conf- ence, which is the 25th in a series of well-established international conferences on security and privacy organized annually by Technical Committee 11 (TC-11) of IFIP. Moreover, in 2010 it is part of the IFIP World Computer Congress 2010 celebrating both the Golden Jubilee of IFIP (founded in 1960) and the Silver Jubilee of the SEC conference in the exciting city of Brisbane, Australia, during September 20-23. The call for papers went out with the challenging motto of "Security & Privacy Silver Linings in the Cloud" building a bridge between the long standing issues of security and privacy and the most recent developments in information and commu- cation technology. It attracted 102 submissions. All of them were evaluated on the basis of their significance, novelty, and technical quality by at least five member of the Program Committee. The Program Committee meeting was held electronically over a period of a week. Of the papers submitted, 25 were selected for presentation at the conference; the acceptance rate was therefore as low as 24. 5% making SEC 2010 a highly competitive forum. One of those 25 submissions could unfortunately not be included in the proceedings, as none of its authors registered in time to present the paper at the conference.
Today, computers fulfil a dazzling array of roles, a flexibility resulting from the great range of programs that can be run on them. "A Science of Operations" examines the history of what we now call programming, defined not simply as "computer" programming, but more broadly as the definition of the steps involved in computations and other information-processing activities. This unique perspective highlights how the history of programming is distinct from the history of the computer, despite the close relationship between the two in the 20th century. The book also discusses how the development of programming languages is related to disparate fields which attempted to give a mechanical account of language on the one hand, and a linguistic account of machines on the other. Topics and features: Covers the early development of automatic computing, including Babbage's "mechanical calculating engines" and the applications of punched-card technology, examines the theoretical work of mathematical logicians such as Kleene, Church, Post and Turing, and the machines built by Zuse and Aiken in the 1930s and 1940s, discusses the role that logic played in the development of the stored program computer, describes the "standard model" of machine-code programming popularised by Maurice Wilkes, presents the complete table for the universal Turing machine in the Appendices, investigates the rise of the initiatives aimed at developing higher-level programming notations, and how these came to be thought of as 'languages' that could be studied independently of a machine, examines the importance of the Algol 60 language, and the framework it provided for studying the design of programming languages and the process of software development and explores the early development of object-oriented languages, with a focus on the Smalltalk project. This fascinating text offers a new viewpoint for historians of science and technology, as well as for the general reader. The historical narrative builds the story in a clear and logical fashion, roughly following chronological order.
After twenty years, return to the wild world of Crash Bandicoot with this epic behind-the-scenes look at the new Crash Bandicoot 4: It's About Time! The Crash Bandicoot (TM) series has remained a beloved staple of platform gaming ever since the first game's release in 1996, The Art of Crash Bandicoot (TM)4: It's About Time (TM) is a rich repository overflowing with interviews, quotes, observations and anecdotes, accompanied by a treasure trove of concept art detailing the characters and environments of the game. Gamers of every type will cherish this all-encompassing look into the zany, wild and unpredictable world of Crash Bandicoot (TM).
.NET represents a new and improved way of developing software for the Windows platform. Given the chance, youd probably rewrite all of your existing code in the newer managed code environment that .NET provides. But it is difficult or impossible to throw out all existing legacy code and start over when a new technology arrives. Instead, you need to find a way to move forward with new .NET development while reusing existing pieces of tested, working code. You need a way to interoperate with the existing code until you have a chance to finally rewrite all of it in .NET. The only recipe-style book on the subject,,"NET 2.0 Interoperability Recipes: A Problem-Solution Approach" guides Windows developers who are transitioning from native Windows code to .NET managed code. The book Explains new interop features in .NET 2.0 and VS .NET 2005 Covers PInvoke, COM, and COM+ (other books dont cover all three areas) Features most of its example code in C# and VB .NET, and also includes some managed C++/CLI Is written by a working developer with first-hand experience .NET tools will allow you to interoperate with existing code. But finding the appropriate tool for the task at hand can sometimes be a frustrating experience. So this book will guide you past myriad infrequently used interop options to focus on those youll use most often.
This volume describes and analyzes in a systematic way the great contributions of the philosopher Krister Segerberg to the study of real and doxastic actions. Following an introduction which functions as a roadmap to Segerberg's works on actions, the first part of the book covers relations between actions, intentions and routines, dynamic logic as a theory of action, agency, and deontic logics built upon the logics of actions. The second section explores belief revision and update, iterated and irrevocable beliefs change, dynamic doxastic logic and hypertheories. Segerberg has worked for more than thirty years to analyze the intricacies of real and doxastic actions using formal tools - mostly modal (dynamic) logic and its semantics. He has had such a significant impact on modal logic that "It is hard to roam for long in modal logic without finding Krister Segerberg's traces," as Johan van Benthem notes in his chapter of this book.
Process Technology brings together in one place important contributions and up-to-date research results in this fast moving area. Process Technology serves as an excellent reference, providing insight into some of the most challenging research issues in the field.
Part I of this book is a practical introduction to working with the Isabelle proof assistant. It teaches you how to write functional programs and inductive definitions and how to prove properties about them in Isabelle's structured proof language. Part II is an introduction to the semantics of imperative languages with an emphasis on applications like compilers and program analysers. The distinguishing feature is that all the mathematics has been formalised in Isabelle and much of it is executable. Part I focusses on the details of proofs in Isabelle; Part II can be read even without familiarity with Isabelle's proof language, all proofs are described in detail but informally. The book teaches the reader the art of precise logical reasoning and the practical use of a proof assistant as a surgical tool for formal proofs about computer science artefacts. In this sense it represents a formal approach to computer science, not just semantics. The Isabelle formalisation, including the proofs and accompanying slides, are freely available online, and the book is suitable for graduate students, advanced undergraduate students, and researchers in theoretical computer science and logic.
Web Dynpro ABAP, a NetWeaver web application user interface tool from SAP, enables web programming connected to SAP Systems. The authors' main focus was to create a book based on their own practical experience. Each chapter includes examples which lead through the content step-by-step and enable the reader to gradually explore and grasp the Web Dynpro ABAP process. The authors explain in particular how to design Web Dynpro components, the data binding and interface methods, and the view controller methods. They also describe the other SAP NetWeaver Elements (ABAP Dictionary, Authorization) and the integration of the Web Dynpro Application into the SAP NetWeaver Portal. The new edition has been expanded to include chapters on subjects such as POWER Lists; creating the Modal Windows and External Windows; using Web Dynpro application parameters and Shared Objects to communicate between the Web Dynpro ABAP Application and Business Server Pages; and creating multi-language mails using Web Dynpro ABAP. |
![]() ![]() You may like...
Implementing Distributed Systems with…
Markus Aleksy, Axel Korthaus, …
Hardcover
R1,729
Discovery Miles 17 290
Introduction to Video Search Engines
David C Gibbon, Zhu Liu
Hardcover
R1,551
Discovery Miles 15 510
Information Security - Foundations…
Ali Ismail Awad, Michael Fairhurst
Hardcover
UML Diagramming - A Case Study Approach
Suriya Sundaramoorthy
Paperback
R1,608
Discovery Miles 16 080
Computing with Data - An Introduction to…
Guy Lebanon, Mohamed El-Geish
Hardcover
R2,907
Discovery Miles 29 070
Perspectives on Web Services - Applying…
Olaf Zimmermann, Mark Tomlinson, …
Hardcover
Evaluating Websites and Web Services
Denis Yannacopoulos, Panagiotis Manolitzas, …
Hardcover
R5,871
Discovery Miles 58 710
|