![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Computer programming > Algorithms & procedures
This volume directly addresses the complexities involved in data mining and the development of new algorithms, built on an underlying theory consisting of linear and non-linear dynamics, data selection, filtering, and analysis, while including analytical projection and prediction. The results derived from the analysis are then further manipulated such that a visual representation is derived with an accompanying analysis. The book brings very current methods of analysis to the forefront of the discipline, provides researchers and practitioners the mathematical underpinning of the algorithms, and the non-specialist with a visual representation such that a valid understanding of the meaning of the adaptive system can be attained with careful attention to the visual representation. The book presents, as a collection of documents, sophisticated and meaningful methods that can be immediately understood and applied to various other disciplines of research. The content is composed of chapters addressing: An application of adaptive systems methodology in the field of post-radiation treatment involving brain volume differences in children; A new adaptive system for computer-aided diagnosis of the characterization of lung nodules; A new method of multi-dimensional scaling with minimal loss of information; A description of the semantics of point spaces with an application on the analysis of terrorist attacks in Afghanistan; The description of a new family of meta-classifiers; A new method of optimal informational sorting; A general method for the unsupervised adaptive classification for learning; and the presentation of two new theories, one in target diffusion and the other in twisting theory.
This volume is about "Structure." The search for "structure," always the pursuit of sciences within their specific areas and perspectives, is witnessing these days a dra matic revolution. The coexistence and interaction of so many structures (atoms, hu mans, cosmos and all that there is in between) would be unconceivable according to many experts, if there were not, behind it all, some gen eral organizational principle. s that (at least in some asymptotic way) make possible so many equilibria among species and natural objects, fan tastically tuned to an extremely high degree of precision. The evidence accumulates to an increasingly impressive degree; a concrete example comes from physics, whose constant aim always was and is that of searching for "ultimate laws," out of which everything should follow, from quarks to the cosmos. Our notions and philosophy have un dergone major revolutions, whenever the "unthinkable" has been changed by its wonderful endeavours into "fact." Well, it is just from physics that evidence comes: even if the "ultimate" could be reached, it would not in any way be a terminal point. When "complexity" comes into the game, entirely new notions have to be invented; they all have to do with "structure," though this time in a much wider sense than would have been understood a decade or so ago."
The cryptosystems based on the Integer Factorization Problem (IFP), the Discrete Logarithm Problem (DLP) and the Elliptic Curve Discrete Logarithm Problem (ECDLP) are essentially the only three types of practical public-key cryptosystems in use. The security of these cryptosystems relies heavily on these three infeasible problems, as no polynomial-time algorithms exist for them so far. However, polynomial-time quantum algorithms for IFP, DLP and ECDLP do exist, provided that a practical quantum computer exists. "Quantum Attacks on Public-Key Cryptosystems" presemts almost allknown quantum computing based attacks on public-key cryptosystems, with an emphasis on quantum algorithms for IFP, DLP, and ECDLP. It also discusses some quantum resistant cryptosystems to replace the IFP, DLP and ECDLP based cryptosystems. This book is intended to be used either as a graduate text in computing, communications and mathematics, or as a basic reference in the field.
Markov decision process (MDP) models are widely used for modeling
sequential decision-making problems that arise in engineering,
economics, computer science, and the social sciences. Many
real-world problems modeled by MDPs have huge state and/or action
spaces, giving an opening to the curse of dimensionality and so
making practical solution of the resulting models intractable. In
other cases, the system of interest is too complex to allow
explicit specification of some of the MDP model parameters, but
simulation samples are readily available (e.g., for random
transitions and costs). For these settings, various sampling and
population-based algorithms have been developed to overcome the
difficulties of computing an optimal solution in terms of a policy
and/or value function. Specific approaches include adaptive
sampling, evolutionary policy iteration, evolutionary random policy
search, and model reference adaptive search.
Computersystemsresearch is heavilyinfluencedby changesincomputertechnol- ogy. As technology changes alterthe characteristics ofthe underlying hardware com- ponents of the system, the algorithms used to manage the system need to be re- examinedand newtechniques need to bedeveloped. Technological influencesare par- ticularly evident in the design of storage management systems such as disk storage managers and file systems. The influences have been so pronounced that techniques developed as recently as ten years ago are being made obsolete. The basic problem for disk storage managers is the unbalanced scaling of hard- warecomponenttechnologies. Disk storage managerdesign depends on the technolo- gy for processors, main memory, and magnetic disks. During the 1980s, processors and main memories benefited from the rapid improvements in semiconductortechnol- ogy and improved by several orders ofmagnitude in performance and capacity. This improvement has not been matched by disk technology, which is bounded by the me- chanics ofrotating magnetic media. Magnetic disks ofthe 1980s have improved by a factor of 10in capacity butonly a factor of2 in performance. This unbalanced scaling ofthe hardware components challenges the disk storage manager to compensate for the slower disks and allow performance to scale with the processor and main memory technology. Unless the performance of file systems can be improved over that of the disks, I/O-bound applications will be unable to use the rapid improvements in processor speeds to improve performance for computer users. Disk storage managers must break this bottleneck and decouple application perfor- mance from the disk.
Genetic Programming Theory and Practice VII presents the results of the annual Genetic Programming Theory and Practice Workshop, contributed by the foremost international researchers and practitioners in the GP arena. Contributions examine the similarities and differences between theoretical and empirical results on real-world problems, and explore the synergy between theory and practice, producing a comprehensive view of the state of the art in GP application. Application areas include chemical process control, circuit design, financial data mining and bio-informatics, to name a few. About this book: Discusses the hurdles encountered when solving large-scale, cutting-edge applications, provides in-depth presentations of the latest and most significant applications of GP and the most recent theoretical results with direct applicability to state-of-the-art problems. Genetic Programming Theory and Practice VII is suitable for researchers, practitioners and students of Genetic Programming, including industry technical staffs, technical consultants and business entrepreneurs.
This volume contains the proceedings of the IFIPTM 2008, the Joint iTrust and PST Conferences on Privacy, Trust Management and Security, held in Trondheim, Norway from June 18 to June 20, 2008. IFIPTM 2008 provides a truly global platform for the reporting of research, development, policy and practice in the interdependent areas of Privacy, Security, and Trust. Following the traditions inherited from the highly successful iTrust and PST conference series, IFIPTM 2008 focuses on trust, privacy and security from multidisciplinary perspectives. The conference is an arena for discussion about re levant problems from both research and practice in the areas of academia, busi ness, and government. IFIPTM 2008 is an open IFIP conference, which only accepts contributed pa pers, so all papers in these proceedings have passed strict peer review. The pro gram of the conference features both theoretical research papers and reports of real world case studies. IFIPTM 2008 received 62 submissions. The program commit tee selected 22 papers for presentation and inclusion in the proceedings. In addi tion, the program and the proceedings include 3 demo descriptions. The highlights of IFIPTM 2008 include invited talks and tutorials by industri al and academic experts in the fields of trust management, privacy and security, including Jon Bing and Michael Steiner.
Details robustness, stability, and performance of Evolutionary Algorithms in dynamic environments
These contributions, written by the foremost international researchers and practitioners of Genetic Programming (GP), explore the synergy between theoretical and empirical results on real-world problems, producing a comprehensive view of the state of the art in GP. Topics in this volume include: exploiting subprograms in genetic programming, schema frequencies in GP, Accessible AI, GP for Big Data, lexicase selection, symbolic regression techniques, co-evolution of GP and LCS, and applying ecological principles to GP. It also covers several chapters on best practices and lessons learned from hands-on experience. Readers will discover large-scale, real-world applications of GP to a variety of problem domains via in-depth presentations of the latest and most significant results.
This book introduces the concepts, applications and development of data science in the telecommunications industry by focusing on advanced machine learning and data mining methodologies in the wireless networks domain. Mining Over Air describes the problems and their solutions for wireless network performance and quality, device quality readiness and returns analytics, wireless resource usage profiling, network traffic anomaly detection, intelligence-based self-organizing networks, telecom marketing, social influence, and other important applications in the telecom industry. Written by authors who study big data analytics in wireless networks and telecommunication markets from both industrial and academic perspectives, the book targets the pain points in telecommunication networks and markets through big data. Designed for both practitioners and researchers, the book explores the intersection between the development of new engineering technology and uses data from the industry to understand consumer behavior. It combines engineering savvy with insights about human behavior. Engineers will understand how the data generated from the technology can be used to understand the consumer behavior and social scientists will get a better understanding of the data generation process.
A best-seller in its French edition, the construction of this book is original and its success in the French market demonstrates its appeal. It is based on three principles: 1. An organization of the chapters by families of algorithms : exhaustive search, divide and conquer, etc. At the contrary, there is no chapter only devoted to a systematic exposure of, say, algorithms on strings. Some of these will be found in different chapters. 2. For each family of algorithms, an introduction is given to the mathematical principles and the issues of a rigorous design, with one or two pedagogical examples. 3. For its most part, the book details 150 problems, spanning on seven families of algorithms. For each problem, a precise and progressive statement is given. More important, a complete solution is detailed, with respect to the design principles that have been presented ; often, some classical errors are pointed at. Roughly speaking, two thirds of the book are devoted to the detailed rational construction of the solutions.
A state-of-the-art research monograph providing consistent treatment of supervisory control, by one of the world 's leading groups in the area of Bayesian identification, control, and decision making. An accompanying CD illustrates the book 's underlying theory.
Synthesis and Optimization of DSP Algorithms describes approaches taken to synthesising structural hardware descriptions of digital circuits from high-level descriptions of Digital Signal Processing (DSP) algorithms. The book contains: -A tutorial on the subjects of digital design and architectural
synthesis, intended for DSP engineers,
This book is an up-to-date documentation of the state of the art in combinatorial optimization, presenting approximate solutions of virtually all relevant classes of NP-hard optimization problems. The well-structured wealth of problems, algorithms, results, and techniques introduced systematically will make the book an indispensible source of reference for professionals. The smooth integration of numerous illustrations, examples, and exercises make this monograph an ideal textbook.
This book not only provides a comprehensive introduction to neural-based PCA methods in control science, but also presents many novel PCA algorithms and their extensions and generalizations, e.g., dual purpose, coupled PCA, GED, neural based SVD algorithms, etc. It also discusses in detail various analysis methods for the convergence, stabilizing, self-stabilizing property of algorithms, and introduces the deterministic discrete-time systems method to analyze the convergence of PCA/MCA algorithms. Readers should be familiar with numerical analysis and the fundamentals of statistics, such as the basics of least squares and stochastic algorithms. Although it focuses on neural networks, the book only presents their learning law, which is simply an iterative algorithm. Therefore, no a priori knowledge of neural networks is required. This book will be of interest and serve as a reference source to researchers and students in applied mathematics, statistics, engineering, and other related fields.
For the introductory Data Structures course (CS2) that typically follows a first course in programming. This text continues to offer a thorough, well-organized, and up-to-date presentation of essential principles and practices in data structures using C++. Reflecting the newest trends in computer science, new and revised material throughout the Second Edition places increased emphasis on abstract data types (ADTs) and object-oriented design. \ To access the author's Companion Website, including Solutions Manual, for ADTS, Data Structures and Problem Solving with C++, please go to http://cs.calvin.edu/books/c++/ds/2e/ For other books by Larry Nyhoff, please go to www.prenhall.com/nyhoff
There has been continuing interest in the improvement of the speed of Digital Signal processing. The use of Residue Number Systems for the design of DSP systems has been extensively researched in literature. Szabo and Tanaka have popularized this approach through their book published in 1967. Subsequently, Jenkins and Leon have rekindled the interest of researchers in this area in 1978, from which time there have been several efforts to use RNS in practical system implementation. An IEEE Press book has been published in 1986 which was a collection of Papers. It is very interesting to note that in the recent past since 1988, the research activity has received a new thrust with emphasis on VLSI design using non ROM based designs as well as ROM based designs as evidenced by the increased publications in this area. The main advantage in using RNS is that several small word-length Processors are used to perform operations such as addition, multiplication and accumulation, subtraction, thus needing less instruction execution time than that needed in conventional 16 bitl32 bit DSPs. However, the disadvantages of RNS have b. een the difficulty of detection of overflow, sign detection, comparison of two numbers, scaling, and division by arbitrary number, RNS to Binary conversion and Binary to RNS conversion. These operations, unfortunately, are computationally intensive and are time consuming."
This monograph gives a thorough treatment of the celebrated compositions of signature and encryption that allow for verifiability, that is, to efficiently prove properties about the encrypted data. This study is provided in the context of two cryptographic primitives: (1) designated confirmer signatures, an opaque signature which was introduced to control the proliferation of certified copies of documents, and (2) signcryption, a primitive that offers privacy and authenticity at once in an efficient way. This book is a useful resource to researchers in cryptology and information security, graduate and PhD students, and security professionals.
This book contains extended and revised versions of the best papers presented at the 17th IFIP WG 10.5/IEEE International Conference on Very Large Scale Integration, VLSI-SoC 2009, held in Florian polis, Brazil, in October 2009. The 8 papers included in the book together with two keynote talks were carefully reviewed and selected from 27 papers presented at the conference. The papers cover a wide variety of excellence in VLSI technology and advanced research addressing the current trend toward increasing chip integration and technology process advancements bringing about stimulating new challenges both at the physical and system-design levels, as well as in the test of theses systems.
Critical Infrastructure Protection II describes original research results and innovative applications in the interdisciplinary field of critical infrastructure protection. Also, it highlights the importance of weaving science, technology and policy in crafting sophisticated solutions that will help secure information, computer and network assets in the various critical infrastructure sectors. This book is the second volume in the annual series produced by the International Federation for Information Processing (IFIP) Working Group 11.10 on Critical Infrastructure Protection, an international community of scientists, engineers, practitioners and policy makers dedicated to advancing research, development and implementation efforts focused on infrastructure protection. The book contains a selection of twenty edited papers from the Second Annual IFIP WG 11.10 International Conference on Critical Infrastructure Protection held at George Mason University, Arlington, Virginia, USA in the spring of 2008.
This book aims to present the impact of Artificial Intelligence (AI) and Big Data in healthcare for medical decision making and data analysis in myriad fields including Radiology, Radiomics, Radiogenomics, Oncology, Pharmacology, COVID-19 prognosis, Cardiac imaging, Neuroradiology, Psychiatry and others. This will include topics such as Artificial Intelligence of Thing (AIOT), Explainable Artificial Intelligence (XAI), Distributed learning, Blockchain of Internet of Things (BIOT), Cybersecurity, and Internet of (Medical) Things (IoTs). Healthcare providers will learn how to leverage Big Data analytics and AI as methodology for accurate analysis based on their clinical data repositories and clinical decision support. The capacity to recognize patterns and transform large amounts of data into usable information for precision medicine assists healthcare professionals in achieving these objectives. Intelligent Health has the potential to monitor patients at risk with underlying conditions and track their progress during therapy. Some of the greatest challenges in using these technologies are based on legal and ethical concerns of using medical data and adequately representing and servicing disparate patient populations. One major potential benefit of this technology is to make health systems more sustainable and standardized. Privacy and data security, establishing protocols, appropriate governance, and improving technologies will be among the crucial priorities for Digital Transformation in Healthcare.
"Examines classic algorithms, geometric diagrams, and mechanical principles for enhances visualization of statistical estimation procedures and mathematical concepts in physics, engineering, and computer programming."
This text explains the fundamental principles of algorithms available for performing arithmetic operations on digital computers. These include basic arithmetic operations like addition, subtraction, multiplication, and division in fixed-point and floating-point number systems as well as more complex operations such as square root extraction and evaluation of exponential, logarithmic, and trigonometric functions. The algorithms described are independent of the particular technology employed for their implementation.
An invaluable resource for working programmers, as well as a fount of useful algorithmic tools for computer scientists, astronomers, and other calendar enthusiasts, The Ultimate Edition updates and expands the previous edition to achieve more accurate results and present new calendar variants. The book now includes coverage of Unix dates, Italian time, the Akan, Icelandic, Saudi Arabian Umm al-Qura, and Babylonian calendars. There are also expanded treatments of the observational Islamic and Hebrew calendars and brief discussions of the Samaritan and Nepalese calendars. Several of the astronomical functions have been rewritten to produce more accurate results and to include calculations of moonrise and moonset. The authors frame the calendars of the world in a completely algorithmic form, allowing easy conversion among these calendars and the determination of secular and religious holidays. LISP code for all the algorithms is available in machine-readable form. |
You may like...
Artificial Intelligence for Signal…
Abhinav Sharma, Arpit Jain, …
Hardcover
R4,233
Discovery Miles 42 330
|