![]() |
![]() |
Your cart is empty |
||
Books > Professional & Technical > Technology: general issues > Technical design > Computer aided design (CAD)
In recent years, both Networks-on-Chip, as an architectural solution for high-speed interconnect, and power consumption, as a key design constraint, have continued to gain interest in the design and research communities. This book offers a single-source reference to some of the most important design techniques proposed in the context of low-power design for networks-on-chip architectures.
This monograph is motivated by the challenges faced in designing reliable VLSI systems in modern VLSI processes. The reliable operation of integrated circuits (ICs) has become increasingly dif?cult to achieve in the deep submicron (DSM) era. With continuouslydecreasing device feature sizes, combinedwith lower supply voltages and higher operating frequencies, the noise immunity of VLSI circuits is decreasing alarmingly. Thus, VLSI circuits are becoming more vulnerable to noise effects such as crosstalk, power supply variations, and radiation-inducedsoft errors. Among these noise sources, soft errors(or error caused by radiation particle strikes) have become an increasingly troublesome issue for memory arrays as well as c- binational logic circuits. Also, in the DSM era, process variations are increasing at a signi?cant rate, making it more dif?cult to design reliable VLSI circuits. Hence, it is important to ef?ciently design robust VLSI circuits that are resilient to radiation particle strikes and process variations. The work presented in this research mo- graph presents several analysis and design techniques with the goal of realizing VLSI circuits, which are radiation and process variation tolerant.
Power consumption of VLSI (Very Large Scale Integrated) circuits has been growing at an alarmingly rapid rate. This increase in power consumption, coupled with the increasing demand for portable/hand-held electronics, has made power consumption a dominant concern in the design of VLSI circuits today. Traditionally, dynamic (switching) power has dominated the total power consumption of an IC. However, due to current scaling trends, leakage power has now become a major component of the total power consumption in VLSI circuits. Leakage power reduction is especially important in portable/hand-held electronics such as cell-phones and PDAs. This book presents two techniques aimed at reducing leakage power in digital VLSI ICs. The first technique reduces leakage through the selective use of high threshold voltage sleep transistors. The second technique reduces leakage by applying the optimal Reverse Body Bias (RBB) voltage. This book also shows readers how to turn the leakage problem into an opportunity, through the use of sub-threshold logic.
The Verilog hardware description language (HDL) provides the ability to describe digital and analog systems. This ability spans the range from descriptions that express conceptual and architectural design to detailed descriptions of implementations in gates and transistors. Verilog was developed originally at Gateway Design Automation Corporation during the mid-eighties. Tools to verify designs expressed in Verilog were implemented at the same time and marketed. Now Verilog is an open standard of IEEE with the number 1364. Verilog HDL is now used universally for digital designs in ASIC, FPGA, microprocessor, DSP and many other kinds of design-centers and is supported by most of the EDA companies. The research and education that is conducted in many universities is also using Verilog. This book introduces the Verilog hardware description language and describes it in a comprehensive manner. Verilog HDL was originally developed and specified with the intent of use with a simulator. Semantics of the language had not been fully described until now. In this book, each feature of the language is described using semantic introduction, syntax and examples. Chapter 4 leads to the full semantics of the language by providing definitions of terms, and explaining data structures and algorithms. The book is written with the approach that Verilog is not only a simulation or synthesis language, or a formal method of describing design, but a complete language addressing all of these aspects. This book covers many aspects of Verilog HDL that are essential parts of any design process.
This book to offers a hands-on guide to designing, analyzing and debugging a communication infrastructure based on the Controller Area Network (CAN) bus. Although the CAN bus standard is well established and currently used in most automotive systems, as well as avionics, medical systems and other devices, its features are not fully understood by most developers, who tend to misuse the network. This results in lost opportunities for better efficiency and performance. These authors offer a comprehensive range of architectural solutions and domains of analysis. It also provides formal models and analytical results, with thorough discussion of their applicability, so that it serves as an invaluable reference for researchers and students, as well as practicing engineers.
Computing increasingly happens somewhere, with that geographic location important to the computational process itself. Many new and evolving spatial technologies, such as geosensor networks and smartphones, embody this trend. Conventional approaches to spatial computing are centralized, and do not account for the inherently decentralized nature of "computing somewhere": the limited, local knowledge of individual system components, and the interaction between those components at different locations. On the other hand, despite being an established topic in distributed systems, decentralized computing is not concerned with geographical constraints to the generation and movement of information. In this context, of (centralized) spatial computing and decentralized (non-spatial) computing, the key question becomes: "What makes decentralized spatial computing special?" In Part I of the book the author covers the foundational concepts, structures, and design techniques for decentralized computing with spatial and spatiotemporal information. In Part II he applies those concepts and techniques to the development of algorithms for decentralized spatial computing, stepping through a suite of increasingly sophisticated algorithms: from algorithms with minimal spatial information about their neighborhoods; to algorithms with access to more detailed spatial information, such as direction, distance, or coordinate location; to truly spatiotemporal algorithms that monitor environments that are dynamic, even using networks that are mobile or volatile. Finally, in Part III the author shows how decentralized spatial and spatiotemporal algorithms designed using the techniques explored in Part II can be simulated and tested. In particular, he investigates empirically the important properties of a decentralized spatial algorithm: its computational efficiency and its robustness to unavoidable uncertainty. Part III concludes with a survey of the opportunities for connecting decentralized spatial computing to ongoing research and emerging hot topics in related fields, such as biologically inspired computing, geovisualization, and stream computing. The book is written for students and researchers of computer science and geographic information science. Throughout the book the author's style is characterized by a focus on the broader message, explaining the process of decentralized spatial algorithm design rather than the technical details. Each chapter ends with review questions designed to test the reader's understanding of the material and to point to further work or research. The book includes short appendices on discrete mathematics and SQL. Simulation models written in NetLogo and associated source code for all the algorithms presented in the book can be found on the author's accompanying website.
Internationally refereed papers present the state of the art in computer-aided architectural design research. These papers reflect the theme of the 12th International Conference of CAADFutures, Integrating Technologies for Computer-Aided Design. Collectively, they provide the technological foundation for new ways of thinking about using computers to design. In addition, they address the education of designers themselves.
This book presents an overview of the issues related to the test, diagnosis and fault-tolerance of Network on Chip-based systems. It is the first book dedicated to the quality aspects of NoC-based systems and will serve as an invaluable reference to the problems, challenges, solutions, and trade-offs related to designing and implementing state-of-the-art, on-chip communication architectures.
Presenting innovative research methods, this second edition of a bestseller describes a simple and practical methodology for conducting cutting-edge design science research (DSR). It provides comprehensive guidance on how to conduct such research and supplies in-depth treatment of design science theory and the different types of theory that can be generated in design science research. Making novel use of the concept of patterns, it presents 84 research patterns for conducting effective DSR. It emphasizes design science theory throughout and is filled with practical examples of using patterns to conduct information and communication technology research (ICT). With a focus on reusing research activities to increase the effectiveness and efficiency of conducting design science research, the book relies on familiar patterns to provide the fundamentals of various research philosophies and techniques required to innovate ICT. It describes design science research in relation to other information systems research paradigms such as positivist and interpretivist research. New to this edition are relevant design science research patterns adapted from TRIZ, the widely regarded European engineering design and creativity method. This edition also provides greatly expanded treatment of theory building in design science research (DSR), a topic of rapidly growing interest in addition to a new chapter presenting a framework for theory development in DSR. The book provides an expanded examination of patterns in DSR presented using a new pattern classification mechanism to group patterns with like functionality. This book will be of value to those interested in learning to conduct design science research, particularly in the ICT disciplines the book focuses on.
Uncertainty in key parameters within a chip and between different chips in the deep sub micron area plays a more and more important role. As a result, manufacturing process spreads need to be considered during the design process. Quantitative methodology is needed to ensure faultless functionality, despite existing process variations within given bounds, during product development. This book presents the technological, physical, and mathematical fundamentals for a design paradigm shift, from a deterministic process to a probability-orientated design process for microelectronic circuits. Readers will learn to evaluate the different sources of variations in the design flow in order to establish different design variants, while applying appropriate methods and tools to evaluate and optimize their design.
Simulation of computer architectures has made rapid progress recently. The primary application areas are hardware/software performance estimation and optimization as well as functional and timing verification. Recent, innovative technologies such as retargetable simulator generation, dynamic binary translation, or sampling simulation have enabled widespread use of processor and system-on-chip (SoC) simulation tools in the semiconductor and embedded system industries. Simultaneously, processor and SoC simulation is still a very active research area, e.g. what amounts to higher simulation speed, flexibility, and accuracy/speed trade-offs. This book presents and discusses the principle technologies and state-of-the-art in high-level hardware architecture simulation, both at the processor and the system-on-chip level.
This book provides readers with a comprehensive review of the state of the art in error control for Network on Chip (NOC) links. Coverage includes detailed description of key issues in NOC error control faced by circuit and system designers, as well as practical error control techniques to minimize the impact of these errors on system performance.
This highly practical Guide to Geometric Algebra in Practice reviews algebraic techniques for geometrical problems in computer science and engineering, and the relationships between them. The topics covered range from powerful new theoretical developments, to successful applications, and the development of new software and hardware tools. Topics and features: provides hands-on review exercises throughout the book, together with helpful chapter summaries; presents a concise introductory tutorial to conformal geometric algebra (CGA) in the appendices; examines the application of CGA for the description of rigid body motion, interpolation and tracking, and image processing; reviews the employment of GA in theorem proving and combinatorics; discusses the geometric algebra of lines, lower-dimensional algebras, and other alternatives to 5-dimensional CGA; proposes applications of coordinate-free methods of GA for differential geometry.
This concise reference helps readers avoid the most commonplace errors in generating or interpreting engineering drawings. Applicable across multiple disciplines, Hanifan's lucid treatment of such essential skills as understanding and conveying data in a drawing, exacting precision in dimension and tolerance notations, and selecting the most-appropriate drawing type for a particular engineering situation, "Perfecting Engineering and Technical Drawing" is an valuable resource for practicing engineers, engineering technologists, and students. Provides straightforward explanation of the requirements for all common engineering drawing types Maximizes reader understanding of engineering drawing requirements, differentiating the types of drawings and their particular characteristics Elucidates electrical reference designation requirements, geometric dimensioning, and tolerancing errors Explains the entire engineering documentation process from concept to delivery
This book provides an in-depth overview of on chip instrumentation technologies and various approaches taken in adding instrumentation to System on Chip (ASIC, ASSP, FPGA, etc.) design that are collectively becoming known as Design for Debug (DfD). On chip instruments are hardware based blocks that are added to a design for the specific purpose and improving the visibility of internal or embedded portions of the design (specific instruction flow in a processor, bus transaction in an on chip bus as examples) to improve the analysis or optimization capabilities for a SoC. DfD is the methodology and infrastructure that surrounds the instrumentation. Coverage includes specific design examples and discussion of implementations and DfD tradeoffs in a decision to design or select instrumentation or SoC that include instrumentation. Although the focus will be on hardware implementations, software and tools will be discussed in some detail.
This book collects together several of the tutorials held at EUROGRAPHICS'89 in Hamburg. The conference was held under the motto "Integration, Visualisation, Interaction" and the tutorials reflect the conference theme. The Springer series EurographicSeminars with the volumes "Advances in Computer Graphics" regularly provides a professional update on current mainstream topics in the field. These publications give readers the opportunity to inform themselves thoroughly on the topics covered. The success of the series is mainly based on the expertise of the contributing authors, who are recognized professionals in their field. Starting out with one of the conference's main topics, the chapter "Visualization of Scientific Data" gives an overview of methods for displaying scientific results in an easily surveyable and comprehensible form. It presents algorithms and methods utilized to achieve visualization results in a form adequate for humans. User interfaces for such systems are also explored, and practical conclusions are drawn. The chapter "Color in Computer Graphics" describes the problems of manipulating and matching color in the real world. After some fundamental statements about color models and their relationships, the main emphasis is placed on the problem of objective color specification for computer graphics systems. It is very hard to match colors between devices such as scanners, printers and displays. Some suggestions on the effective use of color for graphics are also made.
ThIS IS an English verSIOn of the book m two volumes, entitled "KeiJo Shon Kogaku (1), (2)" (Nikkan Kogyo Shinbun Co.) written in Japanese. The purpose of the book is a umfied and systematic exposition of the wealth of research results m the field of mathematical representation of curves and surfaces for computer aided geometric design that have appeared in the last thirty years. The material for the book started hfe as a set of notes for computer aided geometnc design courses which I had at the graduate schools of both computer SCIence, the umversity of Utah m U.S.A. and Kyushu Institute of Design in Japan. The book has been used extensively as a standard text book of curves and surfaces for students, practtcal engmeers and researchers. With the aim of systematic expositIOn, the author has arranged the book in 8 chapters: Chapter 0: The sIgmficance of mathemattcal representations of curves and surfaces is explained and histoncal research developments in this field are revIewed. Chapter 1: BasIc mathematical theones of curves and surfaces are reviewed and summanzed. Chapter 2: A classical mterpolation method, the Lagrange interpolation, is discussed. Although its use is uncommon in practice, this chapter is helpful in understanding Chaps. 4 and 6. Chapter 3: This chapter dIscusses the Coons surface in detail, which is one of the most important contributions in this field. Chapter 4: The fundamentals of spline functions, spline curves and surfaces are discussed in some detail.
This book reviews the algorithms for processing geometric data, with a practical focus on important techniques not covered by traditional courses on computer vision and computer graphics. Features: presents an overview of the underlying mathematical theory, covering vector spaces, metric space, affine spaces, differential geometry, and finite difference methods for derivatives and differential equations; reviews geometry representations, including polygonal meshes, splines, and subdivision surfaces; examines techniques for computing curvature from polygonal meshes; describes algorithms for mesh smoothing, mesh parametrization, and mesh optimization and simplification; discusses point location databases and convex hulls of point sets; investigates the reconstruction of triangle meshes from point clouds, including methods for registration of point clouds and surface reconstruction; provides additional material at a supplementary website; includes self-study exercises throughout the text.
The Future of Design Methodology gives a holistic overview of perspectives for design methodology, addresses trends for developing a powerful methodical support for design practice and provides a starting point for future design research. The chapters are written by leading scientists from around the world, who have great expertise in design methodology, as well as the farsightedness needed to develop design methodology further. The Future of Design Methodology is a detailed contribution to consolidated design methodology and design research. Instead of articulating the views of one scientist, it provides a comprehensive collection of perspectives and visions. The editor highlights the substantial deficiencies and problems of the current design methodology and summarizes the authors' findings to draw future-oriented conclusions. The comprehensive overview of the status of design methodology given in The Future of Design Methodology will help enhance the individual scientific development of junior researchers, while the authoritative perspectives on future design methodology will challenge the views of experts. It is suitable for readers working in a wide range of design fields, such as design methodology, engineering design and industrial design.
SystemC provides a robust set of extensions to the C++ language that enables rapid development of complex models of hardware and software systems. The authors focus on practical use of the language for modeling real systems, showing: A step-by-step build-up of syntax Code examples for each concept Over 8000 lines of downloadable code examples Updates to reflect the SystemC standard, IEEE 1666 Why features are as they are Many resource references How SystemC fits into an ESL methodology This new edition of an industry best seller is updated to reflect the standardization of SystemC as IEEE 1666 and other improvements that reflect feedback from readers of the first edition. The wide ranging feedback also include suggestions from editors of the Japanese and Korean language translations, professors and students, and computer engineers from a broad industrial and geographical spectrum, all who have successfully used the first edition. New chapters have been added on the SystemC Verification Library and the Transaction Level Modeling, and proposed changes to the current SystemC standard. David Black and Jack Donovan, well known consultants in the EDA industry, have teamed with Bill Bunton and Anna Keist, experienced SystemC modeling engineers, to write the second edition of this highly popular classic. As a team the authors bring over 100 years of ASIC and system design experience together to make a very readable introduction to SystemC.
Many problems in scientific computing are intractable with classical numerical techniques. These fail, for example, in the solution of high-dimensional models due to the exponential increase of the number of degrees of freedom. Recently, the authors of this book and their collaborators have developed a novel technique, called Proper Generalized Decomposition (PGD) that has proven to be a significant step forward. The PGD builds by means of a successive enrichment strategy a numerical approximation of the unknown fields in a separated form. Although first introduced and successfully demonstrated in the context of high-dimensional problems, the PGD allows for a completely new approach for addressing more standard problems in science and engineering. Indeed, many challenging problems can be efficiently cast into a multi-dimensional framework, thus opening entirely new solution strategies in the PGD framework. For instance, the material parameters and boundary conditions appearing in a particular mathematical model can be regarded as extra-coordinates of the problem in addition to the usual coordinates such as space and time. In the PGD framework, this enriched model is solved only once to yield a parametric solution that includes all particular solutions for specific values of the parameters. The PGD has now attracted the attention of a large number of research groups worldwide. The present text is the first available book describing the PGD. It provides a very readable and practical introduction that allows the reader to quickly grasp the main features of the method.Throughout the book, the PGD is applied to problems of increasing complexity, and the methodology is illustrated by means of carefully selected numerical examples. Moreover, the reader has free access to the Matlab(c) software used to generate these examples."
Modern electronics depend on nanoscaled technologies that present new challenges in terms of testing and diagnostics. Memories are particularly prone to defects since they exploit the technology limits to get the highest density. This book is an invaluable guide to the testing and diagnostics of the latest generation of SRAM, one of the most widely applied types of memory. Classical methods for testing memory are designed to handle the so-called "static faults," but these test solutions are not sufficient for faults that are emerging in the latest Very Deep Sub-Micron (VDSM) technologies. These new fault models, referred to as "dynamic faults", are not covered by classical test solutions and require the dedicated test sequences presented in this book.
The International PROLAMAT Conference is an internationally well known event for demonstrating and evaluating activities and progress in the field of discrete manufacturing. Sponsored by the International Federation for Information Processing (IFIP), the PROLAMAT is traditionally held every three years and it includes the whole area of advanced software technology for Design and Manufacturing in Discrete Manufacturing. Past editions of the International PROLAMA T Conference have explored: -Manufacturing Technology, -Advances in CAD/CAM, -Software for Discrete Manufacturing, -Software for Manufacturing. The Eight International PROLAMAT held in 1992 (Tokyo), focused on the theme of Man in CIM. The 1995 PROLAMAT (Berlin), featured the theme of Life Cycle Modelling for Innovative Products and Processes. This past emphasis on human aspects and innovation provides a strong foundation for the next PROLAMAT. Under the title: The globalization of manufacturing in the digital communications era of the 21th century: innovation, agility and the virtual enterprise, the 1998 conference expands the PROLAMAT scope to include teams and virtual enterprises which come together across space and time to develop new products and bring them to global markets. Manufacturing issues and information models have long been part of concurrent engineering; they are increasingly important in new product innovation and in the development of manufacturing plans and processes which span multiple companies along with multiple time zones.
Ontologies are increasingly recognized as essential tools in information science. Although the concepts are well understood theoretically , the practical implementation of ontologies remains challenging. In this book, researchers in computer science, information systems, ontology engineering, urban planning and design, civil and building engineering, and architecture present an interdisciplinary study of ontology engineering and its application in urban development projects. The first part of the book introduces the general notion of ontology, describing variations in abstraction level, coverage, and formality. It also discusses the use of ontologies to achieve interoperability, and to represent multiple points of view and multilingualism. This is illustrated with examples from the urban domain. The second part is specific to urban development. It covers spatial and geographical knowledge representation, the creation of urban ontologies from various knowledge sources, the interconnection of urban models and the interaction between standards and domain models. The third part presents case studies of the development of ontologies for urban mobility, urban morphological processes, road systems, and cultural heritage. Other cases report on the use of ontologies to solve urban development problems, in construction business models, building regulations and urban regeneration. It concludes with a discussion of key challenges for the future deployment of ontologies in this domain. This book bridges the gap between urban practitioners and computer scientists. As the essence of most urban projects lies in making connections between worldviews, ontology development has an important role to play, in promoting interoperability between data sources, both formal (urban databases, Building Integrated Models, Geographical Information Systems etc.) and less formal (thesauri, text records, web sources etc.). This volume offers a comprehensive introduction to ontology engineering for urban development. It is essential reading for practitioners and ontology designers working in urban development. |
![]() ![]() You may like...
SolidWorks 2022 Black Book (Colored)
Gaurav Verma, Matt Weber
Hardcover
R2,066
Discovery Miles 20 660
Recent Trends in Computer-aided…
Saptarshi Chatterjee, Debangshu Dey, …
Paperback
R2,729
Discovery Miles 27 290
Up and Running with AutoCAD 2019 - 2D…
Elliot J. Gindis, Robert C. Kaebisch
Paperback
R1,943
Discovery Miles 19 430
AutoCAD Electrical 2023 Black Book…
Gaurav Verma, Matt Weber
Hardcover
R1,583
Discovery Miles 15 830
Designing Around People - CWUAAT 2016
Pat Langdon, Jonathan Lazar, …
Hardcover
|