![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Computer programming
The growing commercial market of Microwave/ Millimeter wave industry over the past decade has led to the explosion of interests and opportunities for the design and development of microwave components.The design of most microwave components requires the use of commercially available electromagnetic (EM) simulation tools for their analysis. In the design process, the simulations are carried out by varying the design parameters until the desired response is obtained. The optimization of design parameters by manual searching is a cumbersome and time consuming process. Soft computing methods such as Genetic Algorithm (GA), Artificial Neural Network (ANN) and Fuzzy Logic (FL) have been widely used by EM researchers for microwave design since last decade. The aim of these methods is to tolerate imprecision, uncertainty, and approximation to achieve robust and low cost solution in a small time frame. Modeling and optimization are essential parts and powerful tools for the microwave/millimeter wave design. This book deals with the development and use of soft computing methods for tackling challenging design problems in the microwave/millimeter wave domain. The aim in the development of these methods is to obtain the design in small time frame while improving the accuracy of the design for a wide range of applications. To achieve this goal, a few diverse design problems of microwave field, representing varied challenges in the design, such as different microstrip antennas, microwave filters, a microstrip-via and also some critical high power components such as nonlinear tapers and RF-windows are considered as case-study design problems. Different design methodologies are developed for these applications. The presents soft computing methods, their review for microwave/millimeter wave design problems and specific case-study problems to infuse better insight and understanding of the subject.
The information infrastructure---comprising computers, embedded devices, networks and software systems---is vital to day-to-day operations in every sector: information and telecommunications, banking and finance, energy, chemicals and hazardous materials, agriculture, food, water, public health, emergency services, transportation, postal and shipping, government and defense. Global business and industry, governments, indeed society itself, cannot function effectively if major components of the critical information infrastructure are degraded, disabled or destroyed. Critical Infrastructure Protection V describes original research results and innovative applications in the interdisciplinary field of critical infrastructure protection. Also, it highlights the importance of weaving science, technology and policy in crafting sophisticated, yet practical, solutions that will help secure information, computer and network assets in the various critical infrastructure sectors. Areas of coverage include: Themes and Issues, Control Systems Security, Infrastructure Security, and Infrastructure Modeling and Simulation. This book is the 5th volume in the annual series produced by the International Federation for Information Processing (IFIP) Working Group 11.10 on Critical Infrastructure Protection, an international community of scientists, engineers, practitioners and policy makers dedicated to advancing research, development and implementation efforts focused on infrastructure protection. The book contains a selection of 14 edited papers from the 5th Annual IFIP WG 11.10 International Conference on Critical Infrastructure Protection, held at Dartmouth College, Hanover, New Hampshire, USA in the spring of 2011. Critical Infrastructure Protection V is an important resource for researchers, faculty members and graduate students, as well as for policy makers, practitioners and other individuals with interests in homeland security. Jonathan Butts is an Assistant Professor of Computer Science at the Air Force Institute of Technology, Wright-Patterson Air Force Base, Ohio, USA. Sujeet Shenoi is the F.P. Walter Professor of Computer Science at the University of Tulsa, Tulsa, Oklahoma, USA.
For the first time, this up-to-date text combines the main issues of the hardware description language VHDL-AMS aimed at model representation of mixed-signal circuits and systems, characterization methods and tools for the extraction of model parameters, and modelling methodologies for accurate high-level behavioural models.
* Covers all aspects of the data acquisition system from design and
specification to programming, installation and configuration
This book acquaints readers with recent developments in dynamical systems theory and its applications, with a strong focus on the control and estimation of nonlinear systems. Several algorithms are proposed and worked out for a set of model systems, in particular so-called input-affine or bilinear systems, which can serve to approximate a wide class of nonlinear control systems. These can either take the form of state space models or be represented by an input-output equation. The approach taken here further highlights the role of modern mathematical and conceptual tools, including differential algebraic theory, observer design for nonlinear systems and generalized canonical forms.
ISGC 2009, The International Symposium on Grid Computing was held at Academia Sinica, Taipei, Taiwan in April 2009 bringing together prestigious scientists and engineers worldwide to exchange ideas, present challenges/solutions and introduce future development in the field of Grid Computing. Managed Grids and Cloud Systems in the Asia-Pacific Research Community presents the latest achievements in grid technology including Cloud Computing. This volume also covers international projects in Grid Operation, Grid Middleware, E-Science applications, technical developments in grid operations and management, Security and Networking, Digital Library and more. The resources used to support these advances, such as volunteer grids, production managed grids, and cloud systems are discussed in detail. This book is designed for a professional audience composed of grid users, developers and researchers working in the grid computing. Advanced-level students focusing on computer science and engineering will find this book valuable as a reference or secondary text book.
Agile methods have taken software development by storm - but a growing controversy over their true effectiveness now pits proponents against detractors in a heated battle of claims. Here at last is a completely impartial guide that gives project managers both expert objective analysis of Agile software development methods and much-needed tools for evaluating the suitability of Agile methods for their organization. The book reviews the Agile Manifesto and the 12 Agile Principles, and discusses in detail each practice of the six most widely recognized Agile methods. Software developers get clear and unbiased insight into adoption implications, the benefits that may accrue, and the potential pitfalls they face. At its core, this unique assessment resource helps developers systematically address key Agile adoption issues involving organizational culture, customers, staff, tools and processes, and project considerations. A comprehensive start-to-finish worksheet makes it easy for professionals to document observations, draw justifiable conclusions, and market their recommendations to the company. The guide also features expert guidelines for customizing an Agile method, training staff, testing the new method, and evaluating its performance - should readers go the Agile route. Practical ready-to-use tools, including equations and templates in Excel format, are provided in a handy workbook which can be accessed through the author's website.
The 6th ACIS International Conference on Software Engineering, Research, Management and Applications (SERA 2008) was held in Prague in the Czech Republic on August 20 - 22. SERA '08 featured excellent theoretical and practical contributions in the areas of formal methods and tools, requirements engineering, software process models, communication systems and networks, software quality and evaluation, software engineering, networks and mobile computing, parallel/distributed computing, software testing, reuse and metrics, database retrieval, computer security, software architectures and modeling. Our conference officers selected the best 17 papers from those papers accepted for presentation at the conference in order to publish them in this volume. The papers were chosen based on review scores submitted by members or the program committee, and underwent further rounds of rigorous review.
This book is a thoroughly practical way to explore the 8051 and
discover C programming through project work. Through graded
projects, Dogan Ibrahim introduces the reader to the fundamentals
of microelectronics, the 8051 family, programming in C, and the use
of a C compiler. The specific device used for examples is the
AT89C2051 - a small, economical chip with re-writable memory,
readily available from the major component suppliers. Microcontroller Projects in C for the 8051 is an ideal resource
for self-study as well as providing an interesting, enjoyable and
easily mastered alternative to more theoretical textbooks.
The Workshop on the Economics of Information Security was established in 2002 to bring together computer scientists and economists to understand and improve the poor state of information security practice. WEIS was borne out of a realization that security often fails for non-technical reasons. Rather, the incentives of both - fender and attacker must be considered. Earlier workshops have answered questions ranging from?nding optimal levels of security investement to understanding why privacy has been eroded. In the process, WEIS has attracted participation from the diverse?elds such as law, management and psychology. WEIS has now established itself as the leading forum for interdisciplinary scholarship on information security. The eigth installment of the conference returned to the United Kingdom, hosted byUniversityCollegeLondononJune24-25,2009.Approximately100researchers, practitioners and government of?cials from across the globe convened in London to hear presentations from authors of 21 peer-reviewed papers, in addition to a panel and keynote lectures from Hal Varian (Google), Bruce Schneier (BT Co- terpane), Martin Sadler (HP Labs), and Robert Coles (Merrill Lynch). Angela Sasse and David Pym chaired the conference, while Christos Ioannidis and Tyler Moore chaired the program committee.
Tackles a topic in a concise and accessible way that most believe too advanced to pick up easily. Author has over 40 years teaching and industry experience which they utilize in this book. Contains an appendix with extended code and examples of topics discussed in text.
Refinement is one of the cornerstones of the formal approach to software engineering, and its use in various domains has led to research on new applications and generalisation. This book brings together this important research in one volume, with the addition of examples drawn from different application areas. It covers four main themes: Data refinement and its application to Z Generalisations of refinement that change the interface and atomicity of operations Refinement in Object-Z Modelling state and behaviour by combining Object-Z with CSP Refinement in Z and Object-Z: Foundations and Advanced Applications provides an invaluable overview of recent research for academic and industrial researchers, lecturers teaching formal specification and development, industrial practitioners using formal methods in their work, and postgraduate and advanced undergraduate students. This second edition is a comprehensive update to the first and includes the following new material: Early chapters have been extended to also include trace refinement, based directly on partial relations rather than through totalisation Provides an updated discussion on divergence, non-atomic refinements and approximate refinement Includes a discussion of the differing semantics of operations and outputs and how they affect the abstraction of models written using Object-Z and CSP Presents a fuller account of the relationship between relational refinement and various models of refinement in CSP Bibliographic notes at the end of each chapter have been extended with the most up to date citations and research
You might think more than enough design books exist in the programming world already. In fact, there are so many that it makes sense to ask why you would read yet another. Is there really a need for yet another design book? In fact, there is a greater need than ever before, and Practical API Design: Confessions of a Java Framework Architect fills that need! Teaches you how to write an API that will stand the test of Time Written by the designer of the NetBeans API at Sun Based on best practices, scalability, and API design patterns What you'll learn What an API is and what the theories are behind good API design When and why to build an API API design patterns applicable to all programming languages, especially modern, object-oriented languages How to optimize and test APIs Who this book is for This book is recommended to every API architect who prefers a bit more engineering design over a purely artistic one.
This handbook offers comprehensive coverage of recent advancements in Big Data technologies and related paradigms. Chapters are authored by international leading experts in the field, and have been reviewed and revised for maximum reader value. The volume consists of twenty-five chapters organized into four main parts. Part one covers the fundamental concepts of Big Data technologies including data curation mechanisms, data models, storage models, programming models and programming platforms. It also dives into the details of implementing Big SQL query engines and big stream processing systems. Part Two focuses on the semantic aspects of Big Data management including data integration and exploratory ad hoc analysis in addition to structured querying and pattern matching techniques. Part Three presents a comprehensive overview of large scale graph processing. It covers the most recent research in large scale graph processing platforms, introducing several scalable graph querying and mining mechanisms in domains such as social networks. Part Four details novel applications that have been made possible by the rapid emergence of Big Data technologies such as Internet-of-Things (IOT), Cognitive Computing and SCADA Systems. All parts of the book discuss open research problems, including potential opportunities, that have arisen from the rapid progress of Big Data technologies and the associated increasing requirements of application domains. Designed for researchers, IT professionals and graduate students, this book is a timely contribution to the growing Big Data field. Big Data has been recognized as one of leading emerging technologies that will have a major contribution and impact on the various fields of science and varies aspect of the human society over the coming decades. Therefore, the content in this book will be an essential tool to help readers understand the development and future of the field.
This book describes a novel methodology for studying algorithmic skills, intended as cognitive activities related to rule-based symbolic transformation, and argues that some human computational abilities may be interpreted and analyzed as genuine examples of extended cognition. It shows that the performance of these abilities relies not only on innate neurocognitive systems or language-related skills, but also on external tools and general agent-environment interactions. Further, it asserts that a low-level analysis, based on a set of core neurocognitive systems linking numbers and language, is not sufficient to explain some specific forms of high-level numerical skills, like those involved in algorithm execution. To this end, it reports on the design of a cognitive architecture for modeling all the relevant features involved in the execution of algorithmic strategies, including external tools, such as paper and pencils. The first part of the book discusses the philosophical premises for endorsing and justifying a position in philosophy of mind that links a modified form of computationalism with some recent theoretical and scientific developments, like those introduced by the so-called dynamical approach to cognition. The second part is dedicated to the description of a Turing-machine-inspired cognitive architecture, expressly designed to formalize all kinds of algorithmic strategies.
* Includes selection of patterns and anti-patterns to describe ideal environment for success. * Looks in-depth at specific tools, and extensions of these tools. * Focuses on how projects are actually handled in real world-drawing on author's vast field experience. * Includes code examples like NAnt automation tasks, case studies, and facilitation utilities.
Adding internet access to embedded systems opens up a whole new world of capabilities. For example, a remote data logging system could automatically send data via the internet and be reconfigured - such as to log new types of data or to measure at different intervals - by commands sent over the internet from any computer or device with internet access. Embedded internet and internet appliances are the focus of great attention in the computing industry, as they are seen as the future of computing, but the design of such devices presents many technical challenges.;This book describes how to design, build and program embedded systems with internet access, giving special attention to sensors and actuators which gather data for transmission over the internet or execute commands sent by the internet, It shows how to build sensors and control devices that connect to the "tiny internet interface" (TINI) and explains how to write programs that control them in Java. Several design case histories are given, including weather monitoring stations, communications centres, automation systems, and data acquisitions systems. The authors discuss how these technologies work and where to get detailed specifications, and they provide ideas for the reader to pursue beyond the book. The accompanying CD-ROM includes Java source code for all the applications described in the book, and an electronic version of the text.
Following the humbling of the 'dot.coms' it is well implemented
corporate portals that are ushering in a new and prosperous era of
e-business. Corporate Portals Empowered with XML and Web Services
provides decision makers with a clear and concise explanation of
what portals are all about, why you really need a portal strategy,
how you go about implementing one, and the issues you have to
encounter and surmount. Guruge shows how you can successfully use
XML and web services to empower your portals for collaboration,
knowledge management, CRM, ERP and supply chain management.
Conceptual modeling is about describing the semantics of software applications at a high level of abstraction in terms of structure, behavior, and user interaction. Embley and Thalheim start with a manifesto stating that the dream of developing information systems strictly by conceptual modeling as expressed in the phrase the model is the code is becoming reality. The subsequent contributions written by leading researchers in the field support the manifesto's assertions, showing not only how to abstractly model complex information systems but also how to formalize abstract specifications in ways that let developers complete programming tasks within the conceptual model itself. They are grouped into sections on programming with conceptual models, structure modeling, process modeling, user interface modeling, and special challenge areas such as conceptual geometric modeling, information integration, and biological conceptual modeling. The Handbook of Conceptual Modeling collects in a single volume many of the best conceptual-modeling ideas, techniques, and practices as well as the challenges that drive research in the field. Thus it is much more than a traditional handbook for advanced professionals, as it also provides both a firm foundation for the field of conceptual modeling, and points researchers and graduate students towards interesting challenges and paths for how to contribute to this fundamental field of computer science.
The pervasiveness of the Internet has had a significant impact on global politics, economics, and culture. To create a truly effective product in such a saturated digital environment, developers must study what has come before and how they can utilize existing tools to even greater effect. Evaluating Websites and Web Services: Interdisciplinary Perspectives on User Satisfaction explores some of the various approaches to the study and assessment of Internet technologies, providing scholars, researchers, developers, and professionals with critical knowledge and an interdisciplinary perspective on e-services in a variety of functional areas, from government and commerce to social media and education.
Applied Time Series Analysis and Innovative Computing contains the applied time series analysis and innovative computing paradigms, with frontier application studies for the time series problems based on the recent works at the Oxford University Computing Laboratory, University of Oxford, the University of Hong Kong, and the Chinese University of Hong Kong. The monograph was drafted when the author was a post-doctoral fellow in Harvard School of Engineering and Applied Sciences, Harvard University. It provides a systematic introduction to the use of innovative computing paradigms as an investigative tool for applications in time series analysis. Applied Time Series Analysis and Innovative Computing offers the state of art of tremendous advances in applied time series analysis and innovative computing paradigms and also serves as an excellent reference work for researchers and graduate students working on applied time series analysis and innovative computing paradigms.
This volume, Technological Aspects of Mentoring, edited by Frances K. Kochan and Joseph T. Pascarelli, will examine mentoring in the technological age. It will focus upon the impact and use of technology in terms of program development, mentoring roles, problems and solutions and issues to be addressed including confidentiality, ethics, and implications for future practice. The editors will explore the possibilities for tomorrow from the work of today.
"Set Theory for Computing" provides a comprehensive account of set-oriented symbolic manipulation methods suitable for automated reasoning. Its main objective is twofold: 1) to provide a flexible formalization for a variety of set languages, and 2) to clarify the semantics of set constructs firmly established in modern specification languages and in the programming practice. Topics include: semantic unification, decision algorithms, modal logics, declarative programming, tableau-based proof techniques, and theory-based theorem proving. The style of presentation is self-contained, rigorous and accurate. Some familiarity with symbolic logic is helpful but not a requirement. This book is a useful resource for all advanced students, professionals, and researchers in computing sciences, artificial intelligence, automated reasoning, logic, and computational mathematics. It will serve to complement their intuitive understanding of set concepts with the ability to master them by symbolic and logically based algorithmic methods and deductive techniques.
The series Studies in Computational Intelligence (SCI) publishes new developments and advances in the various areas of computational intelligence-quickly and with a high quality. The intent is to cover the theory, applications, and design methods of computational intelligence, as embedded in the fields of engineering, computer science, physics and life science, as well as the methodologies behind them. The series contains monographs, lecture notes and edited volumes in computational intelligence spanning the areas of neural networks, connectionist systems, genetic algorithms, evolutionary computation, artificial intelligence, cellular automata, self-organizing systems, soft computing, fuzzy systems and hybrid intelligent systems. Critical to both contributors and readers are the short publication time and world-wide distribution-this permits a rapid and broad dissemination of research results. The purpose of the 10th International Conference on Software Engineering Research, Management and Applications(SERA 2012) held on May 3- June 1, 2012 in Shanghai, China was to bring together scientists, engineers, computer users, and students to share their experiences and exchange new ideas and research results about all aspects (theory, applications and tools) of Software Engineering Research, Management and Applications, and to discuss the practical challenges encountered along the way and the solutions adopted to solve them. The conference organizers selected 12 outstanding papers from those papers accepted for presentation at the conference in order to publish them in this volume. The papers were chosen based on review scores submitted by members of the program committee, and further rigorous rounds of review."
Convexity of sets in linear spaces, and concavity and convexity of functions, lie at the root of beautiful theoretical results that are at the same time extremely useful in the analysis and solution of optimization problems, including problems of either single objective or multiple objectives. Not all of these results rely necessarily on convexity and concavity; some of the results can guarantee that each local optimum is also a global optimum, giving these methods broader application to a wider class of problems. Hence, the focus of the first part of the book is concerned with several types of generalized convex sets and generalized concave functions. In addition to their applicability to nonconvex optimization, these convex sets and generalized concave functions are used in the book's second part, where decision-making and optimization problems under uncertainty are investigated. Uncertainty in the problem data often cannot be avoided when dealing with practical problems. Errors occur in real-world data for a host of reasons. However, over the last thirty years, the fuzzy set approach has proved to be useful in these situations. It is this approach to optimization under uncertainty that is extensively used and studied in the second part of this book. Typically, the membership functions of fuzzy sets involved in such problems are neither concave nor convex. They are, however, often quasiconcave or concave in some generalized sense. This opens possibilities for application of results on generalized concavity to fuzzy optimization. Despite this obvious relation, applying the interface of these two areas has been limited to date. It is hoped that the combination of ideas and results from the field of generalized concavity on the one hand and fuzzy optimization on the other hand outlined and discussed in Generalized Concavity in Fuzzy Optimization and Decision Analysis will be of interest to both communities. Our aim is to broaden the classes of problems that the combination of these two areas can satisfactorily address and solve. |
![]() ![]() You may like...
Principles of Big Graph: In-depth…
Ripon Patgiri, Ganesh Chandra Deka, …
Hardcover
R4,068
Discovery Miles 40 680
Dark Silicon and Future On-chip Systems…
Suyel Namasudra, Hamid Sarbazi-Azad
Hardcover
R4,084
Discovery Miles 40 840
Essential Java for Scientists and…
Brian Hahn, Katherine Malan
Paperback
R1,296
Discovery Miles 12 960
|