![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Applications of computing > Artificial intelligence > Natural language & machine translation
Internet and web technology penetrates many aspects of our daily life. Its importance as a medium for business transactions will grow exponentially during the next few years. In terms of the involved market volume, the B2B area will hereby be the most interesting area. Also, it will be the place, where the new technology will lead to drastic changes in established customer relationships and business models. In an era where open and flexible electronic commerce provides new types of services to its users, simple 1-1 connections will be replaced by n-m relationships between customers and vendors. This new flexibility in electronic trading will generate serious challenges. The main problem stems from the heterogeneity of information descriptions used by vendors and customers, creating problems in both manual trading and in direct 1-1 electronic trading. In the case of B2B market places, it becomes too serious to be neglected. Product descriptions, catalog formats and business documents are often unstructured and non-standardized. Intelligent solutions that mechanize the structuring, standardizing, aligning, and personalizing process are a key requisite for successfully overcoming the current bottlenecks of B2B electronic commerce while enabling its further growth. Intelligent Information Integration in B2B Electronic Commerce discusses the main problems of information integration in this area and sketches several technological solution paths. Intelligent Information Integration in B2B Electronic Commerce is designed to meet the needs of a professional audience composed of researchers and practitioners in industry and graduate level students in Computer Science.
l This book evolved from the ARCADE evaluation exercise that started in 1995. The project's goal is to evaluate alignment systems for parallel texts, i. e., texts accompanied by their translation. Thirteen teams from various places around the world have participated so far and for the first time, some ten to fifteen years after the first alignment techniques were designed, the community has been able to get a clear picture of the behaviour of alignment systems. Several chapters in this book describe the details of competing systems, and the last chapter is devoted to the description of the evaluation protocol and results. The remaining chapters were especially commissioned from researchers who have been major figures in the field in recent years, in an attempt to address a wide range of topics that describe the state of the art in parallel text processing and use. As I recalled in the introduction, the Rosetta stone won eternal fame as the prototype of parallel texts, but such texts are probably almost as old as the invention of writing. Nowadays, parallel texts are electronic, and they are be coming an increasingly important resource for building the natural language processing tools needed in the "multilingual information society" that is cur rently emerging at an incredible speed. Applications are numerous, and they are expanding every day: multilingual lexicography and terminology, machine and human translation, cross-language information retrieval, language learning, etc."
This edited volume comprises invited chapters that cover five areas of the current and the future development of intelligent systems and information sciences. Half of the chapters were presented as invited talks at the Workshop "Future Directions for Intelligent Systems and Information Sciences" held in Dunedin, New Zealand, 22-23 November 1999 after the International Conference on Neuro-Information Processing (lCONIPI ANZIISI ANNES '99) held in Perth, Australia. In order to make this volume useful for researchers and academics in the broad area of information sciences I invited prominent researchers to submit materials and present their view about future paradigms, future trends and directions. Part I contains chapters on adaptive, evolving, learning systems. These are systems that learn in a life-long, on-line mode and in a changing environment. The first chapter, written by the editor, presents briefly the paradigm of Evolving Connectionist Systems (ECOS) and some of their applications. The chapter by Sung-Bae Cho presents the paradigms of artificial life and evolutionary programming in the context of several applications (mobile robots, adaptive agents of the WWW). The following three chapters written by R.Duro, J.Santos and J.A.Becerra (chapter 3), GCoghill . (chapter 4), Y.Maeda (chapter 5) introduce new techniques for building adaptive, learning robots.
This book brings all the major and frontier topics in the field of document analysis together into a single volume, creating a unique reference source that will be invaluable to a large audience of researchers, lecturers and students working in this field. With chapters written by some of the most distinguished researchers active in this field, this book addresses recent advances in digital document processing research and development.
Modern information systems differ in essence from their predecessors. They support operations at multiple locations and different time zones, are distributed and network-based, and use multidimensional data analysis, data warehousing, knowledge discovery, knowledge management, mobile computing, and other modern information processing methods. This book considers fundamental issues of modern information systems. It discusses query processing, data quality, data mining, knowledge management, mobile computing, software engineering for information systems construction, and other topics. The book presents research results that are not available elsewhere. With more than 40 contributors, it is a solid source of information about the state of the art in the field of databases and information systems. It is intended for researchers, advanced students, and practitioners who are concerned with the development of advanced information systems.
People engage in discourse every day - from writing letters and presenting papers to simple discussions. Yet discourse is a complex and fascinating phenomenon that is not well understood. This volume stems from a multidisciplinary workshop in which eminent scholars in linguistics, sociology and computational linguistics presented various aspects of discourse. The topics treated range from multi-party conversational interactions to deconstructing text from various perspectives, considering topic-focus development and discourse structure, and an empirical study of discourse segmentation. The chapters not only describe each author's favorite burning issue in discourse but also provide a fascinating view of the research methodology and style of argumentation in each field.
The last decade has been one of dramatic progress in the field of Natural Language Processing (NLP). This hitherto largely academic discipline has found itself at the center of an information revolution ushered in by the Internet age, as demand for human-computer communication and informa tion access has exploded. Emerging applications in computer-assisted infor mation production and dissemination, automated understanding of news, understanding of spoken language, and processing of foreign languages have given impetus to research that resulted in a new generation of robust tools, systems, and commercial products. Well-positioned government research funding, particularly in the U. S., has helped to advance the state-of-the art at an unprecedented pace, in no small measure thanks to the rigorous 1 evaluations. This volume focuses on the use of Natural Language Processing in In formation Retrieval (IR), an area of science and technology that deals with cataloging, categorization, classification, and search of large amounts of information, particularly in textual form. An outcome of an information retrieval process is usually a set of documents containing information on a given topic, and may consist of newspaper-like articles, memos, reports of any kind, entire books, as well as annotated image and sound files. Since we assume that the information is primarily encoded as text, IR is also a natural language processing problem: in order to decide if a document is relevant to a given information need, one needs to be able to understand its content."
This Second Edition brings readers thoroughly up to date with the emerging field of text mining, the application of techniques of machine learning in conjunction with natural language processing, information extraction, and algebraic/mathematical approaches to computational information retrieval. The book explores a broad range of issues, ranging from the development of new learning approaches to the parallelization of existing algorithms. Authors highlight open research questions in document categorization, clustering, and trend detection. In addition, the book describes new application problems in areas such as email surveillance and anomaly detection.
Computational models can be found everywhere in present day science and engineering. In providing a logical framework and foundation for the specification and design of specification languages, Raymond Turner uses this framework to introduce and study computable models. In doing so he presents the first systematic attempt to provide computational models with a logical foundation. Computable models have wide-ranging applications from programming language semantics and specification languages, through to knowledge representation languages and formalism for natural language semantics. They are also implicit in computer modelling in many areas of physical and social science. This detailed investigation into the logical foundations of specification and specification languages and their application to the definition of programming languages, coupled with a clear exposition of theories of data and computable models as mathematical notions will be welcomed by researchers and graduate students.
Human culture depends on our ability to disseminate information, and then maintain and access it over time. This book addresses the problems of storing, reading, and using digital data for periods longer than 50 years. They offer concise descriptions of markup and document description languages like TIFF, PDF, HTML, and XML, explain important techniques such as migration and emulation, and present the OAIS (Open Archival Information System) Reference Model.
This book presents recent developments in automatic text analysis. Providing an overview of linguistic modeling, it collects contributions of authors from a multidisciplinary area that focus on the topic of automatic text analysis from different perspectives. It includes chapters on cognitive modeling and visual systems modeling, and contributes to the computational linguistic and information theoretical grounding of automatic text analysis.
Automatic Text Categorization and Clustering are becoming more and more important as the amount of text in electronic format grows and the access to it becomes more necessary and widespread. Well known applications are spam filtering and web search, but a large number of everyday uses exist (intelligent web search, data mining, law enforcement, etc.) Currently, researchers are employing many intelligent techniques for text categorization and clustering, ranging from support vector machines and neural networks to Bayesian inference and algebraic methods, such as Latent Semantic Indexing. This volume offers a wide spectrum of research work developed for intelligent text categorization and clustering. In the following, we give a brief introduction of the chapters that are included in this book.
A big amount of important, 'economically relevant' information, is buried within the huge mass of multimedia documents that correspond to some form of 'narrative' description. Due to the ubiquity of these 'narrative' resources, being able to represent in a general, accurate, and effective way their semantic content - i.e., their key 'meaning' - is then both conceptually relevant and economically important. In this book, we present the main properties of NKRL ('Narrative Knowledge Representation Language'), a language expressly designed for representing, in a standardised way, the 'meaning' of complex multimedia narrative documents. NKRL is a fully implemented language/environment. The software exists in two versions, an ORACLE-supported version and a file-oriented one. Written from a multidisciplinary perspective, this exhaustive description of NKRL and of the associated knowledge representation principles will be an invaluable source of reference for practitioners, researchers, and graduates.
Mathematical Linguistics introduces the mathematical foundations of linguistics to computer scientists, engineers, and mathematicians interested in natural language processing. The book presents linguistics as a cumulative body of knowledge from the ground up: no prior knowledge of linguistics is assumed. As the first textbook of its kind, this book is useful for those in information science and in natural language technologies.
The evolution of technology has set the stage for the rapid growth of the video Web: broadband Internet access is ubiquitous, and streaming media protocols, systems, and encoding standards are mature. In addition to Web video delivery, users can easily contribute content captured on low cost camera phones and other consumer products. The media and entertainment industry no longer views these developments as a threat to their established business practices, but as an opportunity to provide services for more viewers in a wider range of consumption contexts. The emergence of IPTV and mobile video services offers unprecedented access to an ever growing number of broadcast channels and provides the flexibility to deliver new, more personalized video services. Highly capable portable media players allow us to take this personalized content with us, and to consume it even in places where the network does not reach. Video search engines enable users to take advantage of these emerging video resources for a wide variety of applications including entertainment, education and communications. However, the task of information extr- tion from video for retrieval applications is challenging, providing opp- tunities for innovation. This book aims to first describe the current state of video search engine technology and second to inform those with the req- site technical skills of the opportunities to contribute to the development of this field. Today's Web search engines have greatly improved the accessibility and therefore the value of the Web.
Foundations of Voice and Speech Quality Perception starts out with the fundamental question of: "How do listeners perceive voice and speech quality and how can these processes be modeled?" Any quantitative answers require measurements. This is natural for physical quantities but harder to imagine for perceptual measurands. This book approaches the problem by actually identifying major perceptual dimensions of voice and speech quality perception, defining units wherever possible and offering paradigms to position these dimensions into a structural skeleton of perceptual speech and voice quality. The emphasis is placed on voice and speech quality assessment of systems in artificial scenarios. Many scientific fields are involved. This book bridges the gap between two quite diverse fields, engineering and humanities, and establishes the new research area of Voice and Speech Quality Perception.
The ideal of using human language to control machines requires a practical theory of natural language communication that includes grammatical analysis of language signs, plus a model of the cognitive agent, with interfaces for recognition and action, an internal database, and an algorithm for reading content in and out. This book offers a functional framework for theoretical analysis of natural language communication and for practical applications of natural language processing.
This volume contains the papers presented at the 23rd Canadian Conference on Arti?cial Intelligence (AI 2010). The conference was held in Ottawa, Ontario, fromMay31toJune2,2010,andwascollocatedwiththe36thGraphicsInterface Conference(GI2010),andthe7thCanadianConferenceonComputerandRobot Vision (CRV 2010). The Program Committee received 90 submissions for the main conference, AI2010,fromacrossCanadaandaroundtheworld.Eachsubmissionwasreviewed byuptofourreviewers.Forthe?nalconferenceprogramandforinclusioninthese proceedings, 22 regular papers, with allocation of 12 pages each, were selected. Additionally,26 shortpapers,with allocationof 4 pageseach,wereaccepted. The papers from the Graduate Student Symposium are also included in the proceedings:sixoral(fourpages)andsixposter(twopages)presentationpapers. The conference programfeatured three keynote presentations by Dekang Lin (Google Inc.), Guy Lapalme (Universit'edeMontr' eal), and Evangelos Milios (Dalhousie University). The one-page abstracts of their talks are also included in the proceedings. Two pre-conference workshops, each with their own proceedings, were held on May 30, 2010. The Workshop on Intelligent Methods for Protecting Privacy and Con?dentiality in Data was organized by Khaled El Emam and Marina Sokolova. The workshop on Teaching AI in Computing and Information Te- nology (AI-CIT 2010) was organized by Danny Silver, Leila Kosseim, and Sajid Hussain. This conference wouldnot havebeen possible without the hardworkofmany people.WewouldliketothankallProgramCommitteemembersandexternal- viewers for their e?ort in providing high-quality reviews in a timely manner. We thank all the authors of submitted papers for submitting their work,and the - thors of selected papers for their collaboration in preparation of the ?nal copy. ManythankstoEbrahimBagheriandMarinaSokolovafororganizingtheGra- ateStudentSymposium,andchairingtheProgramCommitteeofthesymposium. We are in debt to Andrei Voronkov for developing the EasyChair conference managementsystemandmakingitfreelyavailabletotheacademicworld.Itisan amazinglyelegantand functionalWeb-basedsystem,whichsavedus muchtime.
Collections of digital documents can nowadays be found everywhere in institutions, universities or companies. Examples are Web sites or intranets. But searching them for information can still be painful. Searches often return either large numbers of matches or no suitable matches at all. Such document collections can vary a lot in size and how much structure they carry. What they have in common is that they typically do have some structure and that they cover a limited range of topics. The second point is significantly different from the Web in general. The type of search system that we propose in this book can suggest ways of refining or relaxing the query to assist a user in the search process. In order to suggest sensible query modifications we would need to know what the documents are about. Explicit knowledge about the document collection encoded in some electronic form is what we need. However, typically such knowledge is not available. So we construct it automatically.
Parsing technology is concerned with finding syntactic structure in language. In parsing we have to deal with incomplete and not necessarily accurate formal descriptions of natural languages. Robustness and efficiency are among the main issuesin parsing. Corpora can be used to obtain frequency information about language use. This allows probabilistic parsing, an approach that aims at both robustness and efficiency increase. Approximation techniques, to be applied at the level of language description, parsing strategy, and syntactic representation, have the same objective. Approximation at the level of syntactic representation is also known as underspecification, a traditional technique to deal with syntactic ambiguity. In this book new parsing technologies are collected that aim at attacking the problems of robustness and efficiency by exactly these techniques: the design of probabilistic grammars and efficient probabilistic parsing algorithms, approximation techniques applied to grammars and parsers to increase parsing efficiency, and techniques for underspecification and the integration of semantic information in the syntactic analysis to deal with massive ambiguity. The book gives a state-of-the-art overview of current research and development in parsing technologies. In its chapters we see how probabilistic methods have entered the toolbox of computational linguistics in order to be applied in both parsing theory and parsing practice. The book is both a unique reference for researchers and an introduction to the field for interested graduate students.
A selection of papers presented at the international conference Applied Logic: Logic at Work', held in Amsterdam in December 1992. Nowadays, the term applied logic' has a very wide meaning, as numerous applications of logical methods in computer science, formal linguistics and other fields testify. Such applications are by no means restricted to the use of known logical techniques: at its best, applied logic involves a back-and-forth dialogue between logical theory and the problem domain. The papers focus on the application of logic to the study of natural language, in syntax, semantics and pragmatics, and the effect of these studies on the development of logic. In the last decade, the dynamic nature of natural language has been the most interesting challenge for logicians. Dynamic semantics is here applied to new topics, the dynamic approach is extended to syntax, and several methodological issues in dynamic semantics are systematically investigated. Other methodological issues in the formal studies of natural language are discussed, such as the need for types, modal operators and other logical operators in the formal framework. Further articles address the scope of these methodological issues from other perspectives ranging from cognition to computation. The volume presents papers that are interesting for graduate students and researchers in the field of logic, philosophy of language, formal semantics and pragmatics, and computational linguistics.
Corpus-based methods will be found at the heart of many language and speech processing systems. This book provides an in-depth introduction to these technologies through chapters describing basic statistical modeling techniques for language and speech, the use of Hidden Markov Models in continuous speech recognition, the development of dialogue systems, part-of-speech tagging and partial parsing, data-oriented parsing and n-gram language modeling. The book attempts to give both a clear overview of the main technologies used in language and speech processing, along with sufficient mathematics to understand the underlying principles. There is also an extensive bibliography to enable topics of interest to be pursued further. Overall, we believe that the book will give newcomers a solid introduction to the field and it will give existing practitioners a concise review of the principal technologies used in state-of-the-art language and speech processing systems. Corpus-Based Methods in Language and Speech Processing is an initiative of ELSNET, the European Network in Language and Speech. In its activities, ELSNET attaches great importance to the integration of language and speech, both in research and in education. The need for and the potential of this integration are well demonstrated by this publication.
Over the last few decades, research on handwriting recognition has made impressive progress. The research and development on handwritten word recognition are to a large degree motivated by many application areas, such as automated postal address and code reading, data acquisition in banks, text-voice conversion, security, etc. As the prices of scanners, com puters and handwriting-input devices are falling steadily, we have seen an increased demand for handwriting recognition systems and software pack ages. Some commercial handwriting recognition systems are now available in the market. Current commercial systems have an impressive performance in recognizing machine-printed characters and neatly written texts. For in stance, High-Tech Solutions in Israel has developed several products for container ID recognition, car license plate recognition and package label recognition. Xerox in the U. S. has developed TextBridge for converting hardcopy documents into electronic document files. In spite of the impressive progress, there is still a significant perfor mance gap between the human and the machine in recognizing off-line unconstrained handwritten characters and words. The difficulties encoun tered in recognizing unconstrained handwritings are mainly caused by huge variations in writing styles and the overlapping and the interconnection of neighboring characters. Furthermore, many applications demand very high recognition accuracy and reliability. For example, in the banking sector, although automated teller machines (ATMs) and networked banking sys tems are now widely available, many transactions are still carried out in the form of cheques."
New material treats such contemporary subjects as automatic speech recognition and speaker verification for banking by computer and privileged (medical, military, diplomatic) information and control access. The book also focuses on speech and audio compression for mobile communication and the Internet. The importance of subjective quality criteria is stressed. The book also contains introductions to human monaural and binaural hearing, and the basic concepts of signal analysis. Beyond speech processing, this revised and extended new edition of Computer Speech gives an overview of natural language technology and presents the nuts and bolts of state-of-the-art speech dialogue systems.
Yorick Wilks is a central figure in the fields of Natural Language Processing and Artificial Intelligence. This book celebrates Wilks s career from the perspective of his peers in original chapters each of which analyses an aspect of his work and links it to current thinking in that area. This volume forms a two-part set together with Words and Intelligence I: Selected Works by Yorick Wilks, by the same editors." |
![]() ![]() You may like...
Learning from Science and Technology…
Philip Shapira, Stefan Kuhlmann
Hardcover
R4,372
Discovery Miles 43 720
Current Issues in Monetary Policy in the…
Elias C. Grivoyannis
Hardcover
R2,773
Discovery Miles 27 730
Talking To Strangers - What We Should…
Malcolm Gladwell
Paperback
![]()
|