0
Your cart

Your cart is empty

Browse All Departments
Price
  • R100 - R250 (4)
  • R250 - R500 (21)
  • R500+ (2,648)
  • -
Status
Format
Author / Contributor
Publisher

Books > Computing & IT > Computer hardware & operating systems > Computer architecture & logic design > General

SVA: The Power of Assertions in SystemVerilog (Hardcover, 2nd ed. 2015): Eduard Cerny, Surrendra Dudani, John Havlicek, Dmitry... SVA: The Power of Assertions in SystemVerilog (Hardcover, 2nd ed. 2015)
Eduard Cerny, Surrendra Dudani, John Havlicek, Dmitry Korchemny
R4,350 Discovery Miles 43 500 Ships in 10 - 15 working days

This book is a comprehensive guide to assertion-based verification of hardware designs using System Verilog Assertions (SVA). It enables readers to minimize the cost of verification by using assertion-based techniques in simulation testing, coverage collection and formal analysis. The book provides detailed descriptions of all the language features of SVA, accompanied by step-by-step examples of how to employ them to construct powerful and reusable sets of properties. The book also shows how SVA fits into the broader System Verilog language, demonstrating the ways that assertions can interact with other System Verilog components. The reader new to hardware verification will benefit from general material describing the nature of design models and behaviors, how they are exercised, and the different roles that assertions play. This second edition covers the features introduced by the recent IEEE 1800-2012. System Verilog standard, explaining in detail the new and enhanced assertion constructs. The book makes SVA usable and accessible for hardware designers, verification engineers, formal verification specialists and EDA tool developers. With numerous exercises, ranging in depth and difficulty, the book is also suitable as a text for students.

Advanced Transaction Models and Architectures (Hardcover, 1997 ed.): Sushil Jajodia, Larry Kerschberg Advanced Transaction Models and Architectures (Hardcover, 1997 ed.)
Sushil Jajodia, Larry Kerschberg
R4,226 Discovery Miles 42 260 Ships in 18 - 22 working days

Motivation Modem enterprises rely on database management systems (DBMS) to collect, store and manage corporate data, which is considered a strategic corporate re source. Recently, with the proliferation of personal computers and departmen tal computing, the trend has been towards the decentralization and distribution of the computing infrastructure, with autonomy and responsibility for data now residing at the departmental and workgroup level of the organization. Users want their data delivered to their desktops, allowing them to incor porate data into their personal databases, spreadsheets, word processing doc uments, and most importantly, into their daily tasks and activities. They want to be able to share their information while retaining control over its access and distribution. There are also pressures from corporate leaders who wish to use information technology as a strategic resource in offering specialized value-added services to customers. Database technology is being used to manage the data associated with corporate processes and activities. Increasingly, the data being managed are not simply formatted tables in relational databases, but all types of ob jects, including unstructured text, images, audio, and video. Thus, the database management providers are being asked to extend the capabilities of DBMS to include object-relational models as well as full object-oriented database man agement systems."

Logic and Complexity (Hardcover, 2004 ed.): Richard Lassaigne, Michel De Rougemont Logic and Complexity (Hardcover, 2004 ed.)
Richard Lassaigne, Michel De Rougemont
R4,210 Discovery Miles 42 100 Ships in 18 - 22 working days

Logic and Complexity looks at basic logic as it is used in Computer Science, and provides students with a logical approach to Complexity theory. With plenty of exercises, this book presents classical notions of mathematical logic, such as decidability, completeness and incompleteness, as well as new ideas brought by complexity theory such as NP-completeness, randomness and approximations, providing a better understanding for efficient algorithmic solutions to problems.

Divided into three parts, it covers:

- Model Theory and Recursive Functions - introducing the basic model theory of propositional, 1st order, inductive definitions and 2nd order logic. Recursive functions, Turing computability and decidability are also examined.

- Descriptive Complexity - looking at the relationship between definitions of problems, queries, properties of programs and their computational complexity.

- Approximation - explaining how some optimization problems and counting problems can be approximated according to their logical form.

Logic is important in Computer Science, particularly for verification problems and database query languages such as SQL. Students and researchers in this field will find this book of great interest.

Interaction Between Compilers and Computer Architectures (Hardcover, 2001 ed.): Gyungho Lee, Pen-Chung Yew Interaction Between Compilers and Computer Architectures (Hardcover, 2001 ed.)
Gyungho Lee, Pen-Chung Yew
R2,735 Discovery Miles 27 350 Ships in 18 - 22 working days

Effective compilers allow for a more efficient execution of application programs for a given computer architecture, while well-conceived architectural features can support more effective compiler optimization techniques. A well thought-out strategy of trade-offs between compilers and computer architectures is the key to the successful designing of highly efficient and effective computer systems. From embedded micro-controllers to large-scale multiprocessor systems, it is important to understand the interaction between compilers and computer architectures. The goal of the Annual Workshop on Interaction between Compilers and Computer Architectures (INTERACT) is to promote new ideas and to present recent developments in compiler techniques and computer architectures that enhance each other's capabilities and performance. Interaction Between Compilers and Computer Architectures is an updated and revised volume consisting of seven papers originally presented at the Fifth Workshop on Interaction between Compilers and Computer Architectures (INTERACT-5), which was held in conjunction with the IEEE HPCA-7 in Monterrey, Mexico in 2001. This volume explores recent developments and ideas for better integration of the interaction between compilers and computer architectures in designing modern processors and computer systems. Interaction Between Compilers and Computer Architectures is suitable as a secondary text for a graduate level course, and as a reference for researchers and practitioners in industry.

High Performance Computing Systems and Applications (Hardcover, 2000 ed.): Andrew Pollard, Douglas J.K. Mewhort, Donald F.... High Performance Computing Systems and Applications (Hardcover, 2000 ed.)
Andrew Pollard, Douglas J.K. Mewhort, Donald F. Weaver
R4,348 Discovery Miles 43 480 Ships in 18 - 22 working days

High Performance Computing Systems and Applications contains the fully refereed papers from the 13th Annual Symposium on High Performance Computing, held in Kingston, Canada, in June 1999. This book presents the latest research in HPC architectures, distributed and shared memory performance, algorithms and solvers, with special sessions on atmospheric science, computational chemistry and physics. High Performance Computing Systems and Applications is suitable as a secondary text for graduate level courses, and as a reference for researchers and practitioners in industry.

IP Cores Design from Specifications to Production - Modeling, Verification, Optimization, and Protection (Hardcover, 1st ed.... IP Cores Design from Specifications to Production - Modeling, Verification, Optimization, and Protection (Hardcover, 1st ed. 2016)
Khaled Salah Mohamed
R2,705 R1,804 Discovery Miles 18 040 Save R901 (33%) Ships in 10 - 15 working days

This book describes the life cycle process of IP cores, from specification to production, including IP modeling, verification, optimization, and protection. Various trade-offs in the design process are discussed, including those associated with many of the most common memory cores, controller IPs and system-on-chip (SoC) buses. Readers will also benefit from the author's practical coverage of new verification methodologies. such as bug localization, UVM, and scan-chain. A SoC case study is presented to compare traditional verification with the new verification methodologies. Discusses the entire life cycle process of IP cores, from specification to production, including IP modeling, verification, optimization, and protection; Introduce a deep introduction for Verilog for both implementation and verification point of view. Demonstrates how to use IP in applications such as memory controllers and SoC buses. Describes a new verification methodology called bug localization; Presents a novel scan-chain methodology for RTL debugging; Enables readers to employ UVM methodology in straightforward, practical terms.

Parallel Sparse Direct Solver for Integrated Circuit Simulation (Hardcover, 1st ed. 2017): Xiao-Ming Chen, Yu Wang, Huazhong... Parallel Sparse Direct Solver for Integrated Circuit Simulation (Hardcover, 1st ed. 2017)
Xiao-Ming Chen, Yu Wang, Huazhong Yang
R3,182 Discovery Miles 31 820 Ships in 18 - 22 working days

This book describes algorithmic methods and parallelization techniques to design a parallel sparse direct solver which is specifically targeted at integrated circuit simulation problems. The authors describe a complete flow and detailed parallel algorithms of the sparse direct solver. They also show how to improve the performance by simple but effective numerical techniques. The sparse direct solver techniques described can be applied to any SPICE-like integrated circuit simulator and have been proven to be high-performance in actual circuit simulation. Readers will benefit from the state-of-the-art parallel integrated circuit simulation techniques described in this book, especially the latest parallel sparse matrix solution techniques.

Digital Design and Computer Architecture, RISC-V Edition (Paperback): Sarah L. Harris, David Harris Digital Design and Computer Architecture, RISC-V Edition (Paperback)
Sarah L. Harris, David Harris
R2,364 R2,185 Discovery Miles 21 850 Save R179 (8%) Ships in 10 - 15 working days

The newest addition to the Harris and Harris family of Digital Design and Computer Architecture books, this RISC-V Edition covers the fundamentals of digital logic design and reinforces logic concepts through the design of a RISC-V microprocessor. Combining an engaging and humorous writing style with an updated and hands-on approach to digital design, this book takes the reader from the fundamentals of digital logic to the actual design of a processor. By the end of this book, readers will be able to build their own RISC-V microprocessor and will have a top-to-bottom understanding of how it works. Beginning with digital logic gates and progressing to the design of combinational and sequential circuits, this book uses these fundamental building blocks as the basis for designing a RISC-V processor. SystemVerilog and VHDL are integrated throughout the text in examples illustrating the methods and techniques for CAD-based circuit design. The companion website includes a chapter on I/O systems with practical examples that show how to use SparkFun's RED-V RedBoard to communicate with peripheral devices such as LCDs, Bluetooth radios, and motors. This book will be a valuable resource for students taking a course that combines digital logic and computer architecture or students taking a two-quarter sequence in digital logic and computer organization/architecture.

Adiabatic Logic - Future Trend and System Level Perspective (Hardcover, 2012): Philip Teichmann Adiabatic Logic - Future Trend and System Level Perspective (Hardcover, 2012)
Philip Teichmann
R2,652 Discovery Miles 26 520 Ships in 18 - 22 working days

Adiabatic logic is a potential successor for static CMOS circuit design when it comes to ultra-low-power energy consumption. Future development like the evolutionary shrinking of the minimum feature size as well as revolutionary novel transistor concepts will change the gate level savings gained by adiabatic logic. In addition, the impact of worsening degradation effects has to be considered in the design of adiabatic circuits. The impact of the technology trends on the figures of merit of adiabatic logic, energy saving potential and optimum operating frequency, are investigated, as well as degradation related issues. Adiabatic logic benefits from future devices, is not susceptible to Hot Carrier Injection, and shows less impact of Bias Temperature Instability than static CMOS circuits. Major interest also lies on the efficient generation of the applied power-clock signal. This oscillating power supply can be used to save energy in short idle times by disconnecting circuits. An efficient way to generate the power-clock is by means of the synchronous 2N2P LC oscillator, which is also robust with respect to pattern-induced capacitive variations. An easy to implement but powerful power-clock gating supplement is proposed by gating the synchronization signals. Diverse implementations to shut down the system are presented and rated for their applicability and other aspects like energy reduction capability and data retention. Advantageous usage of adiabatic logic requires compact and efficient arithmetic structures. A broad variety of adder structures and a Coordinate Rotation Digital Computer are compared and rated according to energy consumption and area usage, and the resulting energy saving potential against static CMOS proves the ultra-low-power capability of adiabatic logic. In the end, a new circuit topology has to compete with static CMOS also in productivity. On a 130nm test chip, a large scale test vehicle containing an FIR filter was implemented in adiabatic logic, utilizing a standard, library-based design flow, fabricated, measured and compared to simulations of a static CMOS counterpart, with measured saving factors compliant to the values gained by simulation. This leads to the conclusion that adiabatic logic is ready for productive design due to compatibility not only to CMOS technology, but also to electronic design automation (EDA) tools developed for static CMOS system design.

Design Techniques for Mash Continuous-Time Delta-Sigma Modulators (Hardcover, 1st ed. 2018): Qiyuan Liu, Alexander Edward,... Design Techniques for Mash Continuous-Time Delta-Sigma Modulators (Hardcover, 1st ed. 2018)
Qiyuan Liu, Alexander Edward, Carlos Briseno-Vidrios, Jose Silva-Martinez
R2,662 Discovery Miles 26 620 Ships in 18 - 22 working days

This book describes a circuit architecture for converting real analog signals into a digital format, suitable for digital signal processors. This architecture, referred to as multi-stage noise-shaping (MASH) Continuous-Time Sigma-Delta Modulators (CT- M), has the potential to provide better digital data quality and achieve better data rate conversion with lower power consumption. The authors not only cover MASH continuous-time sigma delta modulator fundamentals, but also provide a literature review that will allow students, professors, and professionals to catch up on the latest developments in related technology.

Integrated Research in GRID Computing - CoreGRID Integration Workshop 2005 (Selected Papers) November 28-30, Pisa, Italy... Integrated Research in GRID Computing - CoreGRID Integration Workshop 2005 (Selected Papers) November 28-30, Pisa, Italy (Hardcover, 2007 ed.)
Sergei Gorlatch, Marco Danelutto
R2,813 Discovery Miles 28 130 Ships in 18 - 22 working days

Integrated Research in Grid Computing presents a selection of the best papers presented at the CoreGRID Integration Workshop (CGIW2005), which took place on November 28-30, 2005 in Pisa, Italy. The aim of CoreGRID is to strengthen and advance scientific and technological excellence in the area of Grid and Peer-to-Peer technologies in order to overcome the current fragmentation and duplication of effort in this area. To achieve this objective, the workshop brought together a critical mass of well-established researchers (including 145 permanent researchers and 171 PhD students) from a number of institutions which have all constructed an ambitious joint program of activities. Priority in the workshop was given to work conducted in Tcollaboration between partners from different research institutions and to promising research proposals that could foster such collaboration in the future.

Crisp and Soft Computing with Hypercubical Calculus - New Approaches to Modeling in Cognitive Science and Technology with... Crisp and Soft Computing with Hypercubical Calculus - New Approaches to Modeling in Cognitive Science and Technology with Parity Logic, Fuzzy Logic, and Evolutionary Computing (Hardcover, 1999 ed.)
Michael Zaus
R4,079 Discovery Miles 40 790 Ships in 18 - 22 working days

In Part I, the impact of an integro-differential operator on parity logic engines (PLEs) as a tool for scientific modeling from scratch is presented. Part II outlines the fuzzy structural modeling approach for building new linear and nonlinear dynamical causal forecasting systems in terms of fuzzy cognitive maps (FCMs). Part III introduces the new type of autogenetic algorithms (AGAs) to the field of evolutionary computing. Altogether, these PLEs, FCMs, and AGAs may serve as conceptual and computational power tools.

Formal Hardware Verification - Methods and Systems in Comparison (Paperback, 1997 ed.): Thomas Kropf Formal Hardware Verification - Methods and Systems in Comparison (Paperback, 1997 ed.)
Thomas Kropf
R1,613 Discovery Miles 16 130 Ships in 18 - 22 working days

This state-of-the-art monograph presents a coherent survey of a variety of methods and systems for formal hardware verification. It emphasizes the presentation of approaches that have matured into tools and systems usable for the actual verification of nontrivial circuits. All in all, the book is a representative and well-structured survey on the success and future potential of formal methods in proving the correctness of circuits. The various chapters describe the respective approaches supplying theoretical foundations as well as taking into account the application viewpoint. By applying all methods and systems presented to the same set of IFIP WG10.5 hardware verification examples, a valuable and fair analysis of the strenghts and weaknesses of the various approaches is given.

Compiling Parallel Loops for High Performance Computers - Partitioning, Data Assignment and Remapping (Hardcover, 1993 ed.):... Compiling Parallel Loops for High Performance Computers - Partitioning, Data Assignment and Remapping (Hardcover, 1993 ed.)
David E. Hudak, Santosh G. Abraham
R2,745 Discovery Miles 27 450 Ships in 18 - 22 working days

The exploitationof parallel processing to improve computing speeds is being examined at virtually all levels of computer science, from the study of parallel algorithms to the development of microarchitectures which employ multiple functional units. The most visible aspect of this interest in parallel processing is the commercially available multiprocessor systems which have appeared in the past decade. Unfortunately, the lack of adequate software support for the development of scientific applications that will run efficiently on multiple processors has stunted the acceptance of such systems. One of the major impediments to achieving high parallel efficiency on many data-parallel scientific applications is communication overhead, which is exemplified by cache coherency traffic and global memory overhead of interprocessors with a logically shared address space and physically distributed memory. Such techniques can be used by scientific application designers seeking to optimize code for a particular high-performance computer. In addition, these techniques can be seen as a necesary step toward developing software to support efficient paralled programs. In multiprocessor sytems with physically distributed memory, reducing communication overhead involves both data partitioning and data placement. Adaptive Data Partitioning (ADP) reduces the execution time of parallel programs by minimizing interprocessor communication for iterative data-parallel loops with near-neighbor communication. Data placement schemes are presented that reduce communication overhead. Under the loop partition specified by ADP, global data is partitioned into classes for each processor, allowing each processor to cachecertain regions of the global data set. In addition, for many scientific applications, peak parallel efficiency is achieved only when machine-specific tradeoffs between load imbalance and communication are evaluated and utilized in choosing the data partition. The techniques in this book evaluate these tradeoffs to generate optimum cyclic partitions for data-parallel loops with either a linearly varying or uniform computational structure and either neighborhood or dimensional multicast communication patterns. This tradeoff is also treated within the CPR (Collective Partitioning and Remapping) algorithm, which partitions a collection of loops with various computational structures and communication patterns. Experiments that demonstrate the advantage of ADP, data placement, cyclic partitioning and CPR were conducted on the Encore Multimax and BBN TC2000 multiprocessors using the ADAPT system, a program partitioner which automatically restructures iterative data-parallel loops. This book serves as an excellent reference and may be used as the text for an advanced course on the subject.

Computing with T.Node Parallel Architecture (Hardcover, 1991 ed.): D. Heidrich, J. C Grossetie Computing with T.Node Parallel Architecture (Hardcover, 1991 ed.)
D. Heidrich, J. C Grossetie
R4,044 Discovery Miles 40 440 Ships in 18 - 22 working days

Parallel processing is seen today as the means to improve the power of computing facilities by breaking the Von Neumann bottleneck of conventional sequential computer architectures. By defining appropriate parallel computation models definite advantages can be obtained. Parallel processing is the center of the research in Europe in the field of Information Processing Systems so the CEC has funded the ESPRIT Supemode project to develop a low cost, high performance, multiprocessor machine. The result of this project is a modular, reconfigurable architecture based on !NMOS transputers: T.Node. This machine can be considered as a research, industrial and commercial success. The CEC has decided to continue to encourage manufacturers as well as research and end-users of transputers by funding other projects in this field. This book presents course papers of the Eurocourse given at the Joint Research Centre in ISPRA (Italy) from the 4th to 8 of November 1991. First we present an overview of various trends in the design of parallel architectures and specially of the T.Node with it's software development environments, new distributed system aspects and also new hardware extensions based on the !NMOS T9000 processor. In a second part, we review some real case applications in the field of image synthesis, image processing, signal processing, terrain modeling, particle physics simulation and also enhanced parallel and distributed numerical methods on T.Node.

Distributed Systems for System Architects (Hardcover, 2001 ed.): Paulo Verissimo, Luis Rodrigues Distributed Systems for System Architects (Hardcover, 2001 ed.)
Paulo Verissimo, Luis Rodrigues
R3,454 Discovery Miles 34 540 Ships in 18 - 22 working days

The primary audience for this book are advanced undergraduate students and graduate students. Computer architecture, as it happened in other fields such as electronics, evolved from the small to the large, that is, it left the realm of low-level hardware constructs, and gained new dimensions, as distributed systems became the keyword for system implementation. As such, the system architect, today, assembles pieces of hardware that are at least as large as a computer or a network router or a LAN hub, and assigns pieces of software that are self-contained, such as client or server programs, Java applets or pro tocol modules, to those hardware components. The freedom she/he now has, is tremendously challenging. The problems alas, have increased too. What was before mastered and tested carefully before a fully-fledged mainframe or a closely-coupled computer cluster came out on the market, is today left to the responsibility of computer engineers and scientists invested in the role of system architects, who fulfil this role on behalf of software vendors and in tegrators, add-value system developers, R&D institutes, and final users. As system complexity, size and diversity grow, so increases the probability of in consistency, unreliability, non responsiveness and insecurity, not to mention the management overhead. What System Architects Need to Know The insight such an architect must have includes but goes well beyond, the functional properties of distributed systems."

Handbook of Signal Processing Systems (Hardcover, Edition.): Shuvra S. Bhattacharyya, Ed F. Deprettere, Rainer Leupers, Jarmo... Handbook of Signal Processing Systems (Hardcover, Edition.)
Shuvra S. Bhattacharyya, Ed F. Deprettere, Rainer Leupers, Jarmo Takala
R5,392 Discovery Miles 53 920 Ships in 18 - 22 working days

It gives me immense pleasure to introduce this timely handbook to the research/- velopment communities in the ?eld of signal processing systems (SPS). This is the ?rst of its kind and represents state-of-the-arts coverage of research in this ?eld. The driving force behind information technologies (IT) hinges critically upon the major advances in both component integration and system integration. The major breakthrough for the former is undoubtedly the invention of IC in the 50's by Jack S. Kilby, the Nobel Prize Laureate in Physics 2000. In an integrated circuit, all components were made of the same semiconductor material. Beginning with the pocket calculator in 1964, there have been many increasingly complex applications followed. In fact, processing gates and memory storage on a chip have since then grown at an exponential rate, following Moore's Law. (Moore himself admitted that Moore's Law had turned out to be more accurate, longer lasting and deeper in impact than he ever imagined. ) With greater device integration, various signal processing systems have been realized for many killer IT applications. Further breakthroughs in computer sciences and Internet technologies have also catalyzed large-scale system integration. All these have led to today's IT revolution which has profound impacts on our lifestyle and overall prospect of humanity. (It is hard to imagine life today without mobiles or Internets ) The success of SPS requires a well-concerted integrated approach from mul- ple disciplines, such as device, design, and application.

The Graph Isomorphism Problem - Its Structural Complexity (Hardcover, 1993 ed.): J. Kobler, U. Schoening, J. Toran The Graph Isomorphism Problem - Its Structural Complexity (Hardcover, 1993 ed.)
J. Kobler, U. Schoening, J. Toran
R2,739 Discovery Miles 27 390 Ships in 18 - 22 working days

Recently, a variety ofresults on the complexitystatusofthegraph isomorphism problem has been obtained. These results belong to the so-called structural part of Complexity Theory. Our idea behind this book is to summarize such results which might otherwise not be easily accessible in the literature, and also, to give the reader an understanding of the aims and topics in Structural Complexity Theory, in general. The text is basically self contained; the only prerequisite for reading it is some elementary knowledge from Complexity Theory and Probability Theory. It can be used to teach a seminar or a monographic graduate course, but also parts of it (especially Chapter 1) provide a source of examples for a standard graduate course on Complexity Theory. Many people have helped us in different ways III the process of writing this book. Especially, we would like to thank V. Arvind, R.V. Book, E. May ordomo, and the referee who gave very constructive comments. This book project was especially made possible by a DAAD grant in the "Acciones In tegrada" program. The third author has been supported by the ESPRIT project ALCOM-II."

Reconfigurable Networks-on-Chip (Hardcover, 2012): Sao-jie Chen, Ying-Cherng Lan, Wen-Chung Tsai, Yu Hen Hu Reconfigurable Networks-on-Chip (Hardcover, 2012)
Sao-jie Chen, Ying-Cherng Lan, Wen-Chung Tsai, Yu Hen Hu
R2,661 Discovery Miles 26 610 Ships in 18 - 22 working days

This book provides a comprehensive survey of recent progress in the design and implementation of Networks-on-Chip. It addresses a wide spectrum of on-chip communication problems, ranging from physical, network, to application layers. Specific topics that are explored in detail include packet routing, resource arbitration, error control/correction, application mapping, and communication scheduling. Additionally, a novel bi-directional communication channel NoC (BiNoC) architecture is described, with detailed explanation.

Written for practicing engineers in need of practical knowledge about the design and implementation of networks-on-chip; Includes tutorial-like details to introduce readers to a diverse range of NoC designs, as well as in-depth analysis for designers with NoC experience to explore advanced issues; Describes a variety of on-chip communication architectures, including a novel bi-directional communication channel NoC.

From the Foreword:

Overall this book shows important advances over the state of the art that will affect future system design as well as R&D in tools and methods for NoC design. It represents an important reference point for both designers and electronic design automation researchers and developers.

--Giovanni De Micheli"

High-/Mixed-Voltage Analog and RF Circuit Techniques for Nanoscale CMOS (Hardcover, 2012): Pui-In Mak, Rui Paulo Martins High-/Mixed-Voltage Analog and RF Circuit Techniques for Nanoscale CMOS (Hardcover, 2012)
Pui-In Mak, Rui Paulo Martins
R2,653 Discovery Miles 26 530 Ships in 18 - 22 working days

This book presents high-/mixed-voltage analog and radio frequency (RF) circuit techniques for developing low-cost multistandard wireless receivers in nm-length CMOS processes. Key benefits of high-/mixed-voltage RF and analog CMOS circuits are explained, state-of-the-art examples are studied, and circuit solutions before and after voltage-conscious design are compared. Three real design examples are included, which demonstrate the feasibility of high-/mixed-voltage circuit techniques. Provides a valuable summary and real case studies of the state-of-the-art in high-/mixed-voltage circuits and systems; Includes novel high-/mixed-voltage analog and RF circuit techniques - from concept to practice; Describes the first high-voltage-enabled mobile-TVRF front-end in 90nm CMOS and the first mixed-voltage full-band mobile-TV Receiver in 65nm CMOS;Demonstrates the feasibility of high-/mixed-voltage circuit techniques with real design examples."

CMOS Circuits for Biological Sensing and Processing (Hardcover, 1st ed. 2018): Srinjoy Mitra, David R. S. Cumming CMOS Circuits for Biological Sensing and Processing (Hardcover, 1st ed. 2018)
Srinjoy Mitra, David R. S. Cumming
R4,434 Discovery Miles 44 340 Ships in 10 - 15 working days

This book provides the most comprehensive and consistent survey of the field of IC design for Biological Sensing and Processing. The authors describe a multitude of applications that require custom CMOS IC design and highlight the techniques in analog and mixed-signal circuit design that potentially can cross boundaries and benefit the very wide community of bio-medical engineers.

Smart Camera Design - Algorithms, Architectures, and Art (Hardcover, 1st ed. 2018): Marilyn Wolf Smart Camera Design - Algorithms, Architectures, and Art (Hardcover, 1st ed. 2018)
Marilyn Wolf
R3,982 Discovery Miles 39 820 Ships in 10 - 15 working days

This book describes the algorithms and computer architectures used to create and analyze photographs in modern digital cameras. It also puts the capabilities of digital cameras into context for applications in art, entertainment, and video analysis. The author discusses the entire range of topics relevant to digital camera design, including image processing, computer vision, image sensors, system-on-chip, and optics, while clearly describing the interactions between design decisions at these different levels of abstraction. Readers will benefit from this comprehensive view of digital camera design, describing the range of algorithms used to compose, enhance, and analyze images, as well as the characteristics of optics, image sensors, and computing platforms that determine the physical limits of image capture and computing. The content is designed to be used by algorithm designers and does not require an extensive background in optics or electronics.

Hardware IP Security and Trust (Hardcover, 1st ed. 2017): Prabhat Mishra, Swarup Bhunia, Mark Tehranipoor Hardware IP Security and Trust (Hardcover, 1st ed. 2017)
Prabhat Mishra, Swarup Bhunia, Mark Tehranipoor
R4,443 Discovery Miles 44 430 Ships in 10 - 15 working days

This book provides an overview of current Intellectual Property (IP) based System-on-Chip (SoC) design methodology and highlights how security of IP can be compromised at various stages in the overall SoC design-fabrication-deployment cycle. Readers will gain a comprehensive understanding of the security vulnerabilities of different types of IPs. This book would enable readers to overcome these vulnerabilities through an efficient combination of proactive countermeasures and design-for-security solutions, as well as a wide variety of IP security and trust assessment and validation techniques. This book serves as a single-source of reference for system designers and practitioners for designing secure, reliable and trustworthy SoCs.

The Microarchitecture of Pipelined and Superscalar Computers (Hardcover, 1999 ed.): Amos R. Omondi The Microarchitecture of Pipelined and Superscalar Computers (Hardcover, 1999 ed.)
Amos R. Omondi
R4,159 Discovery Miles 41 590 Ships in 18 - 22 working days

This book is intended to serve as a textbook for a second course in the im plementation (Le. microarchitecture) of computer architectures. The subject matter covered is the collection of techniques that are used to achieve the highest performance in single-processor machines; these techniques center the exploitation of low-level parallelism (temporal and spatial) in the processing of machine instructions. The target audience consists students in the final year of an undergraduate program or in the first year of a postgraduate program in computer science, computer engineering, or electrical engineering; professional computer designers will also also find the book useful as an introduction to the topics covered. Typically, the author has used the material presented here as the basis of a full-semester undergraduate course or a half-semester post graduate course, with the other half of the latter devoted to multiple-processor machines. The background assumed of the reader is a good first course in computer architecture and implementation - to the level in, say, Computer Organization and Design, by D. Patterson and H. Hennessy - and familiarity with digital-logic design. The book consists of eight chapters: The first chapter is an introduction to all of the main ideas that the following chapters cover in detail: the topics covered are the main forms of pipelining used in high-performance uniprocessors, a taxonomy of the space of pipelined processors, and performance issues. It is also intended that this chapter should be readable as a brief "stand-alone" survey."

Application-Specific Hardware Architecture Design with VHDL (Hardcover, 1st ed. 2018): Bogdan Belean Application-Specific Hardware Architecture Design with VHDL (Hardcover, 1st ed. 2018)
Bogdan Belean
R3,290 Discovery Miles 32 900 Ships in 10 - 15 working days

This book guides readers through the design of hardware architectures using VHDL for digital communication and image processing applications that require performance computing. Further it includes the description of all the VHDL-related notions, such as language, levels of abstraction, combinational vs. sequential logic, structural and behavioral description, digital circuit design, and finite state machines. It also includes numerous examples to make the concepts presented in text more easily understandable.

Free Delivery
Pinterest Twitter Facebook Google+
You may like...
The Ripple Effect - How surprisingly…
David J. Harkin Hardcover R348 Discovery Miles 3 480
Scotland Yard - the Methods and…
George 1883- Dilnot Hardcover R771 Discovery Miles 7 710
The Skilled Helper - A Client-Centred…
Gerard Egan, Robert J. Reese Paperback R1,305 R1,212 Discovery Miles 12 120
Counselling And Coping
Kerry Gibson, Leslie Swartz, … Paperback R458 Discovery Miles 4 580
Improving Psychiatric Care for Older…
Claire Hilton Hardcover R1,195 Discovery Miles 11 950
ENERGY IN MOTION EXPANDING With DNA…
Jan Hein Hardcover R693 R617 Discovery Miles 6 170
Peak - For Fans of Atomic Habits
Anders Ericsson, Robert Pool Paperback  (1)
R295 R272 Discovery Miles 2 720
Systemized Thinking Models for…
Joel E Winston Hardcover R707 Discovery Miles 7 070
Rewire Your Brain - 2 in 1: How To…
Jennifer Ferguson Hardcover R774 R688 Discovery Miles 6 880
Family Law In South Africa
Paperback  (1)
R633 Discovery Miles 6 330

 

Partners