![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Professional & Technical > Technology: general issues > Technical design > Computer aided design (CAD)
Hardware veri?cation is the process of checking whether a design conforms to its speci?cations of functionality and timing. In today's design processes it becomes more and more important. Very large scale integrated (VLSI) circuits and the resulting digital systems have conquered a place in almost all areas of our life, even in security sensitive applications. Complex digital systems control airplanes, have been used in banks and on intensive-care units. Hence, the demand for error-free designs is more important than ever. In addition, economic reasons underline this demand as well. The design and production process of present day VLSI-circuits is highly time- and cost-intensive. Mo- over, it is nearly impossible to repair integrated circuits. Thus, it is desirable to detect design errors early in the design process and not just after producing the prototype chip. All these facts are re?ected by developing and prod- tion statistics of present day companies. For example, In?neon Technologies [118] assumed that about 60% to 80% of the overall design time was spent for veri?cation in 2000. Other sources cite the 3-to-1 head count ratio between veri?cation engineers and logic designers. This shows that verifying logical correctness of the design of hardware systems is a major gate to the problem of time-to-market (cf. [113]). With the chip complexity constantly increasing, the dif?culty as well as the - portance of functional veri?cation of new product designs has been increased. It is not only more important to get error-free designs.
Data Access and Storage Management for Embedded Programmable
Processors gives an overview of the state-of-the-art in
system-level data access and storage management for embedded
programmable processors. The targeted application domain covers
complex embedded real-time multi-media and communication
applications. Many of these applications are data-dominated in the
sense that their cost related aspects, namely power consumption and
footprint are heavily influenced (if not dominated) by the data
access and storage aspects. The material is mainly based on
research at IMEC in this area in the period 1996-2001. In order to
deal with the stringent timing requirements and the data dominated
characteristics of this domain, we have adopted a target
architecture style that is compatible with modern embedded
processors, and we have developed a systematic step-wise
methodology to make the exploration and optimization of such
applications feasible in a source-to-source precompilation
approach.
This book is the first in a series of three dedicated to advanced topics in Mixed-Signal IC design methodologies. It is one of the results achieved by the Mixed-Signal Design Cluster, an initiative launched in 1998 as part of the TARDIS project, funded by the European Commission within the ESPRIT-IV Framework. This initiative aims to promote the development of new design and test methodologies for Mixed-Signal ICs, and to accelerate their adoption by industrial users. As Microelectronics evolves, Mixed-Signal techniques are gaining a significant importance due to the wide spread of applications where an analog front-end is needed to drive a complex digital-processing subsystem. In this sense, Analog and Mixed-Signal circuits are recognized as a bottleneck for the market acceptance of Systems-On-Chip, because of the inherent difficulties involved in the design and test of these circuits. Specially, problems arising from the use of a common substrate for analog and digital components are a main limiting factor. The Mixed-Signal Cluster has been formed by a group of 11 Research and Development projects, plus a specific action to promote the dissemination of design methodologies, techniques, and supporting tools developed within the Cluster projects. The whole action, ending in July 2002, has been assigned an overall budget of more than 8 million EURO.
This book covers in a great depth the fast growing topic of tools, techniques and applications of soft computing (e.g., fuzzy logic, genetic algorithms, neural networks, rough sets, Bayesian networks, and other probabilistic techniques) in the ontologies and the Semantic Web. The author shows how components of the Semantic Web (like the RDF, Description Logics, ontologies) can be covered with a soft computing methodology.
This volume is dedicated to the memory of Barclay G. Jones, Professor of City and Regional Planning and Regional Science at Cornell University. Over a decade ago, Barclay took on a fledgling area of study - economic modeling of disasters - and nurtured its early development. He served as the social science program director at the National Center for Earthquake Engineering Research (NCEER), a university consortium sponsored by the National Science Foundation and the Federal Emergency Management Agency of the United States. In this capacity, Barclay shepherded and attracted a number of regional scientists to the study of disasters. He organized a conference, held in the ill-fated World Trade Center in September 1995, on "The Economic Consequences of Earthquakes: Preparing for the Unexpected. " He persistently advocated the importance of social science research in an establishment dominated by less-than-sympathetic natural scientists and engineers. In 1993, Barclay organized the first of a series of sessions on "Measuring Regional Economic Effects of Unscheduled Events" at the North American Meetings of the Regional Science Association International (RSAI). This unusual nomenclature brought attention to the challenge that disasters -largely unanticipated, often sudden, and always disorderly - pose to the regional science modeling tradition. The sessions provided an annual forum for a growing coalition of researchers, where previously the literature had been fragmentary, scattered, and episodic. Since Barclay's unexpected passing in 1997, we have continued this effort in his tradition.
This book contains an edited version of the lectures and selected contributions presented during the Advanced Summer Institute (ASI) on "Product Engineering: Tools and Methods based on Virtual Reality" held at Chania (Greece), 30th May - 6th June 2007. The ASI was devoted to the Product Engineering field, with particular attention being given to the aspects related to Virtual Reality (VR) technologies, and their use and added value in engineering.
This edited volume is targeted at presenting the latest state-of-the-art methodologies in "Hybrid Evolutionary Algorithms." The chapters deal with the theoretical and methodological aspects, as well as various applications to many real world problems from science, technology, business or commerce. Overall, the book has 14 chapters including an introductory chapter giving the fundamental definitions and some important research challenges. The contributions were selected on the basis of fundamental ideas/concepts rather than the thoroughness of techniques deployed.
The application of Computational Intelligence in emerging research areas such as Granular Computing, Mechatronics, and Bioinformatics shows its usefulness often emphasized by Prof Lotfi Zadeh, the inventor of fuzzy logic and many others. This book contains recent advances in Computational Intelligence methods for modeling, optimization and prediction and covers a large number of applications. The book presents new Computational Intelligence theory and methods for modeling and prediction. The range of the various applications is captured with 5 chapters in image processing, 2 chapters in audio processing, 3 chapters in commerce and finance, 2 chapters in communication networks and 6 chapters containing other applications.
The Integrated Circuit (IC) industry has gone without a standardized verification approach for decades. This book defines a uniform, standardizable methodology for verifying the logical behavior of an integrated circuit, whether an I/O controller, a microprocessor, or a complete digital system. This book will help Engineers and managers responsible for IC development to bring a single, standards-based methodology to their R & D efforts, cutting costs and improving results.
This book details timing analysis and optimization techniques for circuits with level-sensitive memory elements. It contains a linear programming formulation applicable to the timing analysis of large scale circuits and includes a delay insertion methodology that improves the efficiency of clock skew scheduling. Coverage also provides a framework for and results from implementing timing optimization algorithms in a parallel computing environment.
In its updated second edition, this book has been extensively revised on a chapter by chapter basis. The book accurately reflects the syntax and semantic changes to the SystemVerilog language standard, making it an essential reference for systems professionals who need the latest version information. In addition, the second edition features a new chapter explaining the SystemVerilog "packages," a new appendix that summarizes the synthesis guidelines presented throughout the book, and all of the code examples have been updated to the final syntax and rerun using the latest version of the Synopsys, Mentor, and Cadance tools.
Collaborative Product Design and Manufacturing Methodologies and Applications introduces a wide spectrum of collaborative engineering issues in design and manufacturing. It offers state-of-the-art chapters written by international experts from academia and industry, and reflects the most up-to-date R & D work and applications, especially those from the last three to five years. The book will serve as an essential reference for academics, upper-level undergraduate and graduate students and practicing professionals.
Interconnect has become the dominating factor in determining system performance in nanometer technologies. This book is dedicated to this important subject. The primary purpose of this monograph is to provide insight and intuition into layout analysis and optimization for interconnect in high speed, high complexity integrated circuits. In this monograph, the effects of wire size, spacing between wires, wire length, coupling length, load capacitance, rise time of the inputs, place of overlap (near driver or receiver side), frequency, shields, direction of the signals, and wire width for both the aggressors and the victim wires on system performance and reliability is thoroughly investigated. Also, parameters like driver strength has been considered as several recent studies considered the simultaneous device and interconnect sizing. Crosstalk noise, as well as the impact of coupling on aggressor delay is analyzed. The pulse width of the crosstalk noise, which is of similar importance for circuit performance as the peak amplitude, is also analyzed. We have considered more parameters that can affect the signal integrity and presented practical intensive simulation results. This book brings together a wealth of information previously scattered throughout the literature, presenting a range of CAD algorithms and techniques for synthesizing and optimizing interconnect. The practical aspects of the algorithms and models are explained with sufficient detail. It deeply investigates the most two effective parameters in layout optimization, spacing and shield insertion, that can affect both capacitive and inductive noise. Noise models needed for layouts with multi-layer multi-crosscoupling segments are investigated. Different post-layout optimization techniques are explained with complexity analysis and benchmarks tests are provided.
This book carefully details design tools and techniques for realizing low power and energy efficiency in a highly productive design methodology. Important topics include: Design examples illustrate that these techniques can improve energy efficiency by two to three times.
Evolutionary algorithms are sophisticated search methods that have been found to be very efficient and effective in solving complex real-world multi-objective problems where conventional optimization tools fail to work well. Despite the tremendous amount of work done in the development of these algorithms in the past decade, many researchers assume that the optimization problems are deterministic and uncertainties are rarely examined. The primary motivation of this book is to provide a comprehensive introduction on the design and application of evolutionary algorithms for multi-objective optimization in the presence of uncertainties. In this book, we hope to expose the readers to a range of optimization issues and concepts, and to encourage a greater degree of appreciation of evolutionary computation techniques and the exploration of new ideas that can better handle uncertainties. "Evolutionary Multi-Objective Optimization in Uncertain Environments: Issues and Algorithms" is intended for a wide readership and will be a valuable reference for engineers, researchers, senior undergraduates and graduate students who are interested in the areas of evolutionary multi-objective optimization and uncertainties.
This edited collection of essays from world-leading academic and industrial authors yields insight into all aspects of reverse engineering. Methods of reverse engineering analysis are covered, along with special emphasis on the investigation of surface and internal structures. Frequently-used hardware and software are assessed and advice given on the most suitable choice of system. Also covered is rapid prototyping and its relationship with successful reverse engineering.
Model Predictive Control System Design and Implementation Using MATLAB(r) proposes methods for design and implementation of MPC systems using basis functions that confer the following advantages: - continuous- and discrete-time MPC problems solved in similar design frameworks; - a parsimonious parametric representation of the control trajectory gives rise to computationally efficient algorithms and better on-line performance; and - a more general discrete-time representation of MPC design that becomes identical to the traditional approach for an appropriate choice of parameters. After the theoretical presentation, coverage is given to three industrial applications. The subject of quadratic programming, often associated with the core optimization algorithms of MPC is also introduced and explained. The technical contents of this book is mainly based on advances in MPC using state-space models and basis functions. This volume includes numerous analytical examples and problems and MATLAB(r) programs and exercises.
Soft computing embraces methodologies for the development of intelligent systems that have been successfully applied to a large number of real-word problems. This collection of keynote papers, presented at the 7th On-line World Conference on Soft Computing in Engineering Design and Manufacturing, provides a comprehensive overview of recent advances in fuzzy, neural and evolutionary computing techniques and applications in engineering design and manufacturing. Features: A collection of methodologies aimed at researchers and professional design and manufacturing engineers who develop and apply intelligent systems in computer engineering.
Hugo de Man Professor Katholieke Universiteit Leuven Senior Research Fellow IMEC The steady evolution of hardware, software and communications technology is rapidly transforming the PC- and dot.com world into the world of Ambient Intelligence (AmI). This next wave of information technology is fundam- tally different in that it makes distributed wired and wireless computing and communication disappear to the background and puts users to the foreground. AmI adapts to people instead of the other way around. It will augment our consciousness, monitor our health and security, guide us through traffic etc. In short, its ultimate goal is to improve the quality of our life by a quiet, reliable and secure interaction with our social and material environment. What makes AmI engineering so fascinating is that its design starts from studying person to world interactions that need to be implemented as an int- ligent and autonomous interplay of virtually all necessary networked electronic intelligence on the globe. This is a new and exciting dimension for most elect- cal and software engineers and may attract more creative talent to engineering than pure technology does. Development of the leading technology for AmI will only succeed if the engineering research community is prepared to join forces in order to make Mark Weiser's dream of 1991 come true. This will not be business as usual by just doubling transistor count or clock speed in a microprocessor or increasing the bandwidth of communication.
No other book has been published giving a single-volume introduction and survey to production planning in distributed manufacturing networks. The published literature so far includes conference proceedings only.
Hardware Software Co-Design of a Multimedia SOC Platform is one of the first of its kinds to provide a comprehensive overview of the design and implementation of the hardware and software of an SoC platform for multimedia applications. Topics covered in this book range from system level design methodology, multimedia algorithm implementation, a sub-word parallel, single-instruction-multiple data (SIMD) processor design, and its virtual platform implementation, to the development of an SIMD parallel compiler as well as a real-time operating system (RTOS). Hardware Software Co-Design of a Multimedia SOC Platform is written for practitioner engineers and technical managers who want to gain first hand knowledge about the hardware-software design process of an SoC platform. It offers both tutorial-like details to help readers become familiar with a diverse range of subjects, and in-depth analysis for advanced readers to pursue further.
Covering the development of field computation in the past forty years, this book is a concise, comprehensive and up-to-date introduction to methods for the analysis and synthesis of electric and magnetic fields. A broad view of the subject of field models in electricity and magnetism, ranging from basic theory to numerical applications, is offered. The approach throughout is to solve field problems directly from partial differential equations in terms of vector quantities.
PAAMS, the International Conference on Practical Applications of Agents and Multi-Agent Systems is an evolution of the International Workshop on Practical Applications of Agents and Multi-Agent Systems. PAAMS is an international yearly tribune to present, to discuss, and to disseminate the latest developments and the most important outcomes related to real-world applications. It provides a unique opportunity to bring multi-disciplinary experts, academics and practitioners together to exchange their experience in the development of Agents and Multi-Agent Systems. This volume presents the papers that have been accepted for the 2009 edition. These articles capture the most innovative results and this year's trends: Assisted Cognition, E-Commerce, Grid Computing, Human Modelling, Information Systems, Knowledge Management, Agent-Based Simulation, Software Development, Transports, Trust and Security. Each paper has been reviewed by three different reviewers, from an international committee composed of 64 members from 20 different countries. From the 92 submissions received, 35 were selected for full presentation at the conference, and 26 were accepted as posters.
This volume reflects the theme of the INFORMS 2004 Meeting in Denver: Back to OR Roots. Emerging as a quantitative approach to problem-solving in World War II, our founders were physicists, mathematicians, and engineers who quickly found peace-time uses. It is fair to say that Operations Research (OR) was born in the same incubator as computer science, and it has spawned many new disciplines, such as systems engineering, health care management, and transportation science. Although people from many disciplines routinely use OR methods, many scientific researchers, engineers, and others do not understand basic OR tools and how they can help them. Disciplines ranging from finance to bioengineering are the beneficiaries of what we do - we take an interdisciplinary approach to problem-solving. Our strengths are modeling, analysis, and algorithm design. We provide a quanti- tive foundation for a broad spectrum of problems, from economics to medicine, from environmental control to sports, from e-commerce to computational - ometry. We are both producers and consumers because the mainstream of OR is in the interfaces. As part of this effort to recognize and extend OR roots in future probl- solving, we organized a set of tutorials designed for people who heard of the topic and want to decide whether to learn it. The 90 minutes was spent addre- ing the questions: What is this about, in a nutshell? Why is it important? Where can I learn more? In total, we had 14 tutorials, and eight of them are published here.
For the past decade or so, Computational Intelligence (CI) has been an - tremely "hot" topic amongst researchers working in the ?elds of biomedicine and bioinformatics. There are many successful applications of CI in such areas ascomputationalgenomics, predictionofgeneexpression, proteinstructure, and protein-protein interactions, modeling of evolution, or neuronal systems mod- ing and analysis. However, there still are many problems in biomedicine and bioinformatics that are in desperate need of advanced and e?cient compu- tional methodologies to deal with tremendous amounts of data so prevalent in those kinds of researchpursuits. Unfortunately, scientists in both these ?elds are very often unaware of the abundance of computational techniques that could be put to use to help them analyze and understand the data underlying their research inquiries. On the other hand, computational intelligence practitioners are often unfamiliar with the particular problems that their algorithms could be successfully applied for. The separation between the two worlds is partially caused by the use of di?erent languages in these two spheres of science, but also by a relatively small number of publications devoted solely to the purpose of facilitating the exchange of new computational algorithms and methodologies on one hand, and the needs of the realms of biomedicine and bioinformatics on the other. Inordertohelp?llthegapbetweenthescientistsonbothsidesofthisspectrum, wehavesolicitedcontributionsfromresearchersactivelyapplyingcomputational intelligencetechniquestoimportantproblemsinbiomedicineandbioinformatics. The purpose of this book is to provide an overview of powerful state-of-the-art methodologiesthatarecurrentlyutilizedforbiomedicine-and/orbioinformati- orientedapplications, sothatresearchersworkinginthose?eldscouldlearnofnew methodstohelpthemtackletheirproblems. Ontheotherhand, wealsohopethat the CI community will ?nd this book useful by discovering a new and intriguing area of applications. |
You may like...
Microorganisms in Plant Conservation and…
K. Sivasithamparam, K.W. Dixon, …
Hardcover
R4,221
Discovery Miles 42 210
Canadian Estate Planning Made Easy…
Terrance Hamilton Hall
Hardcover
Parallel Computing in Optimization
A. Migdalas, Panos M. Pardalos, …
Hardcover
R7,963
Discovery Miles 79 630
Transparency in Insurance Regulation and…
Pierpaolo Marano, Kyriaki Noussia
Hardcover
R4,999
Discovery Miles 49 990
Multiscale Modeling of Vascular Dynamics…
Huilin Ye, Zhiqiang Shen, …
Paperback
R750
Discovery Miles 7 500
Electronic Design Automation for…
Luciano Lavagno, Igor L Markov, …
Paperback
R5,189
Discovery Miles 51 890
Nonlinear Assignment Problems…
Panos M. Pardalos, L.S. Pitsoulis
Hardcover
R4,049
Discovery Miles 40 490
|