![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Artificial intelligence > Natural language & machine translation
The digital age has had a profound effect on our cultural
heritage and the academic research that studies it. Staggering
amounts of objects, many of them of a textual nature, are being
digitised to make them more readily accessible to both experts and
laypersons. Besides a vast potential for more effective and
efficient preservation, management, and presentation, digitisation
offers opportunities to work with cultural heritage data in ways
that were never feasible or even imagined.
This book constitutes the refereed proceedings of the 12th China National Conference on Computational Linguistics, CCL 2013, and of the First International Symposium on Natural Language Processing Based on Naturally Annotated Big Data, NLP-NABD 2013, held in Suzhou, China, in October 2013. The 32 papers presented were carefully reviewed and selected from 252 submissions. The papers are organized in topical sections on word segmentation; open-domain question answering; discourse, coreference and pragmatics; statistical and machine learning methods in NLP; semantics; text mining, open-domain information extraction and machine reading of the Web; sentiment analysis, opinion mining and text classification; lexical semantics and ontologies; language resources and annotation; machine translation; speech recognition and synthesis; tagging and chunking; and large-scale knowledge acquisition and reasoning.
This book constitutes the refereed proceedings of the 13th China National Conference on Computational Linguistics, CCL 2014, and of the First International Symposium on Natural Language Processing Based on Naturally Annotated Big Data, NLP-NABD 2014, held in Wuhan, China, in October 2014. The 27 papers presented were carefully reviewed and selected from 233 submissions. The papers are organized in topical sections on word segmentation; syntactic analysis and parsing the Web; semantics; discourse, coreference and pragmatics; textual entailment; language resources and annotation; sentiment analysis, opinion mining and text classification; large-scale knowledge acquisition and reasoning; text mining, open IE and machine reading of the Web; machine translation; multilinguality in NLP; underresourced languages processing; NLP applications.
Two Top Industry Leaders Speak Out Judith Markowitz When Amy asked me to co-author the foreword to her new book on advances in speech recognition, I was honored. Amy's work has always been infused with c- ative intensity, so I knew the book would be as interesting for established speech professionals as for readers new to the speech-processing industry. The fact that I would be writing the foreward with Bill Scholz made the job even more enjoyable. Bill and I have known each other since he was at UNISYS directing projects that had a profound impact on speech-recognition tools and applications. Bill Scholz The opportunity to prepare this foreword with Judith provides me with a rare oppor- nity to collaborate with a seasoned speech professional to identify numerous signi- cant contributions to the field offered by the contributors whom Amy has recruited. Judith and I have had our eyes opened by the ideas and analyses offered by this collection of authors. Speech recognition no longer needs be relegated to the ca- gory of an experimental future technology; it is here today with sufficient capability to address the most challenging of tasks. And the point-click-type approach to GUI control is no longer sufficient, especially in the context of limitations of mode- day hand held devices. Instead, VUI and GUI are being integrated into unified multimodal solutions that are maturing into the fundamental paradigm for comput- human interaction in the future.
This book constitutes the refereed proceedings of the 10th China Workshop on Machine Translation, CWMT 2014, held in Macau, China, in November 2014. The 10 revised full English papers presented were carefully reviewed and selected from 15 submissions of English papers. The papers cover the following topics: machine translation; data selection; word segmentation; entity recognition; MT evaluation.
Never before has a book been published that describes the techniques and technology used in writing text editors, word processors and other software. Written for the working professional and serious student, this book covers all aspects of the task. The topics range from user psychology to selecting a language to implementing redisplay to designing the command set. More than just facts are involved, however, as this book also promotes insight into an understanding of the issues encountered when designing such software. After reading this book, you should have a clear understanding of how to go about writing text editing or word processing software. In addition, this book introduces the concepts and power of the Emacs-type of text editor. This type of editor can trace its roots to the first computer text editor written and is still by far the most powerful editor available.
This book constitutes the refereed proceedings of the 18th International Conference on Applications of Natural Language to Information Systems, held in Salford, UK, in June 2013. The 21 long papers, 15 short papers and 17 poster papers presented in this volume were carefully reviewed and selected from 80 submissions. The papers cover the following topics: requirements engineering, question answering systems, named entity recognition, sentiment analysis and mining, forensic computing, semantic web, and information search.
The practical task of building a talking robot requires a theory of how natural language communication works. Conversely, the best way to computationally verify a theory of natural language communication is to demonstrate its functioning concretely in the form of a talking robot, the epitome of human-machine communication. To build an actual robot requires hardware that provides appropriate recognition and action interfaces, and because such hardware is hard to develop the approach in this book is theoretical: the author presents an artificial cognitive agent with language as a software system called database semantics (DBS). Because a theoretical approach does not have to deal with the technical difficulties of hardware engineering there is no reason to simplify the system - instead the software components of DBS aim at completeness of function and of data coverage in word form recognition, syntactic-semantic interpretation and inferencing, leaving the procedural implementation of elementary concepts for later. In this book the author first examines the universals of natural language and explains the Database Semantics approach. Then in Part I he examines the following natural language communication issues: using external surfaces; the cycle of natural language communication; memory structure; autonomous control; and learning. In Part II he analyzes the coding of content according to the aspects: semantic relations of structure; simultaneous amalgamation of content; graph-theoretical considerations; computing perspective in dialogue; and computing perspective in text. The book ends with a concluding chapter, a bibliography and an index. The book will be of value to researchers, graduate students and engineers in the areas of artificial intelligence and robotics, in particular those who deal with natural language processing.
Electronic Multimedia Publishing brings together in one place important contributions and up-to-date research results in this fast moving area. Electronic Mulitmedia Publishing serves as an excellent reference, providing insight into some of the most challenging research issues in the field.
This textbook is intended for graduate students in computer science and linguistics who are interested in developing expertise in natural language processing (NLP) and in those aspects of artificial intelligence which are concerned with computer models oflanguage comprehension. The text is somewhat different from a number of other excellent textbooks in that its foci are more on the linguistic and psycho linguistic prerequisites and on foundational issues concerning human linguistic behavior than on the description of the extant models and algorithms. The goal is to make the student, undertaking the enormous task of developing computer models for NLP, well aware of the major diffi culties and unsolved problems, so that he or she will not begin the task (as it has often been done) with overoptimistic hopes or claims about the generalizability of models, when such hopes and claims are incon sistent either with some aspects of the formal theory or with known facts about human cognitive behavior. Thus, I try to enumerate and explain the variety of cognitive, linguistic, and pragmatic data which must be understood and formalized before they can be incorporated into a computer model."
Information Highways are widely considered as the next generation of high speed communication systems. These highways will be based on emerging Broadband Integrated Services Digital Networks (B-ISDN), which - at least in principle - are envisioned to support not only all the kinds of networking applications known today but also future applications which are not as yet understood fully or even anticipated. Thus, B-ISDNs release networking processes from the limitations which the communications medium has imposed historically. The operational generality stems from the versatility of Asynchronous Transfer Mode (ATM) which is the transfer mode adopted by ITU-T for broadband public ISDN as well as wide area private ISDN. A transfer mode which provides the transmission, multiplexing and switching core that lies at the foundations of a communication network. ATM is designed to integrate existing and future voice, audio, image and data services. Moreover, ATM aims to minimise the complexity of switching and buffer management, to optimise intermediate node processing and buffering and to bound transmission delays. These design objectives are met at high transmission speeds by keeping the basic unit of ATM transmission - the ATM cell - short and of fixed length.
This book constitutes the refereed proceedings of the 4th International Workshop on Controlled Natural Language, CNL 2014, held in Galway, Ireland, in August 2014. The 17 full papers and one invited paper presented were carefully reviewed and selected from 26 submissions. The topics include simplified language, plain language, formalized language, processable language, fragments of language, phraseologies, conceptual authoring, language generation, and guided natural language interfaces.
This two-volume set, consisting of LNCS 7816 and LNCS 7817, constitutes the thoroughly refereed proceedings of the 13th International Conference on Computer Linguistics and Intelligent Processing, CICLING 2013, held on Samos, Greece, in March 2013. The total of 91 contributions presented was carefully reviewed and selected for inclusion in the proceedings. The papers are organized in topical sections named: general techniques; lexical resources; morphology and tokenization; syntax and named entity recognition; word sense disambiguation and coreference resolution; semantics and discourse; sentiment, polarity, subjectivity, and opinion; machine translation and multilingualism; text mining, information extraction, and information retrieval; text summarization; stylometry and text simplification; and applications.
The Language of Mathematics was awarded the E.W. Beth Dissertation Prize for outstanding dissertations in the fields of logic, language, and information. It innovatively combines techniques from linguistics, philosophy of mathematics, and computation to give the first wide-ranging analysis of mathematical language. It focuses particularly on a method for determining the complete meaning of mathematical texts and on resolving technical deficiencies in all standard accounts of the foundations of mathematics. "The thesis does far more than is required for a PhD: it is more like a lifetime's work packed into three years, and is a truly exceptional achievement." Timothy Gowers
The SGML FAQ Book: Understanding the Foundation of HTML and XML is similar, but not quite the same kind of thing as an online FAQ or 'Frequently Asked Questions' list. It addresses questions from people who already actually use SGML in some way (including HTML authors), and people who are about to use it. It deals mainly with issues that arise when using SGML in practice. A very brief introduction to SGML is included as Appendix A. The questions discussed in The SGML FAQ Book are repeatedly heard by people who make their living serving the SGML community. SGML experts spend many hours teaching these details, sometimes repeatedly because some questions do not seem important - until you run into them. So one benefit of this book is learning more of the art of document creation and management, both by general reading before questions arise and by specific reference when a question arises. For the latter use, the appendices, glossary, and index are particularly important. A second benefit of this book is that it provides a common theme to its answers that you can apply in your use of SGML, HTML and related languages in general.The fundamental answer to many of the questions boils down to 'simplify': many questions do not show up if you use the simple, elegant core of SGML without worrying about optional features. The credo of this book is simply, 'SGML doesn't need to be complicated'. SGML has the potential for complexity at certain points. But much of the complexity comes from optional parts and can be avoided. SGML methodology and its primary benefits suffer no loss even if you skip many features, which speaks well for the quality of SGML's overall design. Many of the questions discussed involve those optional parts, and therefore can be avoided by judicious designers and authors. The two key goals of the book are (1) to answer questions that you may actually encounter as an SGML user, and to help you get 'unstuck' and be as productive as possible in using the language and (2) to show proactive ways you can simplify your use of SGML, and get its very substantial benefits with minimal complexity.
This book constitutes the refereed selected papers from the 14th Chinese Lexical Semantics Workshop, CLSW 2013, held in Zhengzhou, China, in May 2013. The 68 full papers and 4 short papers presented in this volume were carefully reviewed and selected from 153 submissions. They are organized in topical sections covering all major topics of lexical semantics; lexical resources; corpus linguistics and applications on natural language processing.
Recent advances in the fields of knowledge representation, reasoning and human-computer interaction have paved the way for a novel approach to treating and handling context. The field of research presented in this book addresses the problem of contextual computing in artificial intelligence based on the state of the art in knowledge representation and human-computer interaction. The author puts forward a knowledge-based approach for employing high-level context in order to solve some persistent and challenging problems in the chosen showcase domain of natural language understanding. Specifically, the problems addressed concern the handling of noise due to speech recognition errors, semantic ambiguities, and the notorious problem of underspecification. Consequently the book examines the individual contributions of contextual composing for different types of context. Therefore, contextual information stemming from the domain at hand, prior discourse, and the specific user and real world situation are considered and integrated in a formal model that is applied and evaluated employing different multimodal mobile dialog systems. This book is intended to meet the needs of readers from at least three fields - AI and computer science; computational linguistics; and natural language processing - as well as some computationally oriented linguists, making it a valuable resource for scientists, researchers, lecturers, language processing practitioners and professionals as well as postgraduates and some undergraduates in the aforementioned fields. "The book addresses a problem of great and increasing technical and practical importance - the role of context in natural language processing (NLP). It considers the role of context in three important tasks: Automatic Speech Recognition, Semantic Interpretation, and Pragmatic Interpretation. Overall, the book represents a novel and insightful investigation into the potential of contextual information processing in NLP." Jerome A Feldman, Professor of Electrical Engineering and Computer Science, UC Berkeley, USA http://dm.tzi.de/research/contextual-computing/
Computer parsing technology, which breaks down complex linguistic structures into their constituent parts, is a key research area in the automatic processing of human language. This volume is a collection of contributions from leading researchers in the field of natural language processing technology, each of whom detail their recent work which includes new techniques as well as results. The book presents an overview of the state of the art in current research into parsing technologies, focusing on three important themes: dependency parsing, domain adaptation, and deep parsing. The technology, which has a variety of practical uses, is especially concerned with the methods, tools and software that can be used to parse automatically. Applications include extracting information from free text or speech, question answering, speech recognition and comprehension, recommender systems, machine translation, and automatic summarization. New developments in the area of parsing technology are thus widely applicable, and researchers and professionals from a number of fields will find the material here required reading. As well as the other four volumes on parsing technology in this series this book has a breadth of coverage that makes it suitable both as an overview of the field for graduate students, and as a reference for established researchers in computational linguistics, artificial intelligence, computer science, language engineering, information science, and cognitive science. It will also be of interest to designers, developers, and advanced users of natural language processing systems, including applications such as spoken dialogue, text mining, multimodal human-computer interaction, and semantic web technology.
The description, automatic identification and further processing of web genres is a novel field of research in computational linguistics, NLP and related areas such as text-technology, digital humanities and web mining. One of the driving forces behind this research is the idea of genre-enabled search engines which enable users to additionally specify web genres that the documents to be retrieved should comply with (e.g., personal homepage, weblog, scientific article etc.). This book offers a thorough foundation of this upcoming field of research on web genres and document types in web-based social networking. It provides theoretical foundations of web genres, presents corpus linguistic approaches to their analysis and computational models for their classification. This includes research in the areas of web genre identification, web genre modelling and related fields such as genres and registers in web based communication social software-based document networks web genre ontologies and classification schemes text-technological models of web genres web content, structure and usage mining web genre classification web as corpus. The book addresses researchers who want to become acquainted with theoretical developments, computational models and their empirical evaluation in this field of research. It also addresses researchers who are interested in standards for the creation of corpora of web documents. Thus, the book concerns readers from many disciplines such as corpus linguistics, computational linguistics, text-technology and computer science.
Packed with numerous examples this easy-to-use manual provides designers of interactive media with a practical guide to screen design. Written for the working designer the manual shows how to rework information so that it is suitable for the chosen target group and the media they are using. It describes background information and encourages further development of the language of digital media. In doing so it covers all aspects of screen design, including: perception and learning, ergonomics, communication theory, image research, aesthetics, and the design of navigation and orientation elements.
Explorations in Automatic Thesaurus Discovery presents an automated method for creating a first-draft thesaurus from raw text. It describes natural processing steps of tokenization, surface syntactic analysis, and syntactic attribute extraction. From these attributes, word and term similarity is calculated and a thesaurus is created showing important common terms and their relation to each other, common verb--noun pairings, common expressions, and word family members. The techniques are tested on twenty different corpora ranging from baseball newsgroups, assassination archives, medical X-ray reports, abstracts on AIDS, to encyclopedia articles on animals, even on the text of the book itself. The corpora range from 40,000 to 6 million characters of text, and results are presented for each in the Appendix. The methods described in the book have undergone extensive evaluation. Their time and space complexity are shown to be modest. The results are shown to converge to a stable state as the corpus grows. The similarities calculated are compared to those produced by psychological testing. A method of evaluation using Artificial Synonyms is tested. Gold Standards evaluation show that techniques significantly outperform non-linguistic-based techniques for the most important words in corpora. Explorations in Automatic Thesaurus Discovery includes applications to the fields of information retrieval using established testbeds, existing thesaural enrichment, semantic analysis. Also included are applications showing how to create, implement, and test a first-draft thesaurus.
Based on a NATO Advanced Study Institute held in 1993, this book addresses recent advances in automatic speech recognition and speech coding. The book contains contributions by many of the most outstanding researchers from the best laboratories worldwide in the field. The contributions have been grouped into five parts: on acoustic modeling; language modeling; speech processing, analysis and synthesis; speech coding; and vector quantization and neural nets. For each of these topics, some of the best-known researchers were invited to give a lecture. In addition to these lectures, the topics were complemented with discussions and presentations of the work of those attending. Altogether, the reader is given a wide perspective on recent advances in the field and will be able to see the trends for future work.
The ability of storing, managing, and giving access to the huge quantity of data collected by astronomical observatories is one of the major challenges of modern astronomy. At the same time, the growing complexity of data systems implies a change of concepts: the scientist has to manipulate data as well as information. Recent developments of the WorldWideWeb' bring interesting answers to these problems. The book presents a wide selection of databases, archives, data centers, and information systems. Clear and up-to-date descriptions are included, together with their scientific context and motivations. Audience: This volume provides an essential tool for astronomers, librarians, data specialists and computer engineers.
Most of the papers in this volume were first presented at the Workshop on Cross-Linguistic Information Retrieval that was held August 22, 1996 dur ing the SIGIR'96 Conference. Alan Smeaton of Dublin University and Paraic Sheridan of the ETH, Zurich, were the two other members of the Scientific Committee for this workshop. SIGIR is the Association for Computing Ma chinery (ACM) Special Interest Group on Information Retrieval, and they have held conferences yearly since 1977. Three additional papers have been added: Chapter 4 Distributed Cross-Lingual Information retrieval describes the EMIR retrieval system, one of the first general cross-language systems to be implemented and evaluated; Chapter 6 Mapping Vocabularies Using Latent Semantic Indexing, which originally appeared as a technical report in the Lab oratory for Computational Linguistics at Carnegie Mellon University in 1991, is included here because it was one of the earliest, though hard-to-find, publi cations showing the application of Latent Semantic Indexing to the problem of cross-language retrieval; and Chapter 10 A Weighted Boolean Model for Cross Language Text Retrieval describes a recent approach to solving the translation term weighting problem, specific to Cross-Language Information Retrieval. Gregory Grefenstette CONTRIBUTORS Lisa Ballesteros David Hull W, Bruce Croft Gregory Grefenstette Center for Intelligent Xerox Research Centre Europe Information Retrieval Grenoble Laboratory Computer Science Department University of Massachusetts Thomas K. Landauer Department of Psychology Mark W. Davis and Institute of Cognitive Science Computing Research Lab University of Colorado, Boulder New Mexico State University Michael L. Littman Bonnie J.
This volume constitutes the refereed proceedings of the Spanish Conference, IberSPEECH 2012: Joint VII "Jornadas en Tecnologia del Habla" and III Iberian SLTech Workshop, held in Madrid, Spain, in November 21-23, 2012. The 29 revised papers were carefully reviewed and selected from 80 submissions. The papers are organized in topical sections on speaker characterization and recognition; audio and speech segmentation; pathology detection and speech characterization; dialogue and multimodal systems; robustness in automatic speech recognition; applications of speech and language technologies. |
You may like...
From Riemann to Differential Geometry…
Lizhen Ji, Athanase Papadopoulos, …
Hardcover
R4,880
Discovery Miles 48 800
Robustness and Complex Data Structures…
Claudia Becker, Roland Fried, …
Hardcover
New Ecoinformatics Tools in…
Vladimir F. Krapivin, Costas A. Varotsos, …
Hardcover
R4,292
Discovery Miles 42 920
Multivariate Analysis of Ecological Data…
Jean Thioulouse, Stephane Dray, …
Hardcover
R2,467
Discovery Miles 24 670
Geometric Complex Analysis - In Honor of…
Jisoo Byun, Hong Rae Cho, …
Hardcover
R4,061
Discovery Miles 40 610
Theory and Applications of Recent Robust…
Mia Hubert, Greet Pison, …
Hardcover
R2,759
Discovery Miles 27 590
Superschool on Derived Categories and…
Matthew Ballard, Charles Doran, …
Hardcover
R4,713
Discovery Miles 47 130
|