![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Language & Literature > Language & linguistics > Computational linguistics
This volume is a collection of original contributions that address the problem of words and their meaning. This represents a still difficult and controversial area within various disciplines: linguistics, philosophy, and artificial intelligence. Although all of these disciplines have to tackle the issue, so far there is no overarching methodology agreed upon by researchers. The aim of the volume is to provide answers based on empirical linguistics methods that are relevant across all the disciplines and provide a bridge among researchers looking at word meaning from different angles.
This book explains how to build Natural Language Generation (NLG) systems--computer software systems that automatically generate understandable texts in English or other human languages. NLG systems use knowledge about language and the application domain to automatically produce documents, reports, explanations, help messages, and other kinds of texts. The book covers the algorithms and representations needed to perform the core tasks of document planning, microplanning, and surface realization, using a case study to show how these components fit together. It is essential reading for researchers interested in NLP, AI, and HCI; and for developers interested in advanced document-creation technology.
This reader collects and introduces important work in linguistics, computer science, artificial intelligence, and computational linguistics on the use of linguistic devices in natural languages to situate events in time: whether they are past, present, or future; whether they are real or hypothetical; when an event might have occurred, and how long it could have lasted. In focussing on the treatment and retrieval of time-based information it seeks to lay the foundation for temporally-aware natural language computer processing systems, for example those that process documents on the worldwide web to answer questions or produce summaries. The development of such systems requires the application of technical knowledge from many different disciplines. The book is the first to bring these disciplines together, by means of classic and contemporary papers in four areas: tense, aspect, and event structure; temporal reasoning; the temporal structure of natural language discourse; and temporal annotation. Clear, self-contained editorial introductions to each area provide the necessary technical background for the non-specialist, explaining the underlying connections across disciplines. A wide range of students and professionals in academia and industry will value this book as an introduction and guide to a new and vital technology. The former include researchers, students, and teachers of natural language processing, linguistics, artificial intelligence, computational linguistics, computer science, information retrieval (including the growing speciality of question-answering), library sciences, human-computer interaction, and cognitive science. Those in industry include corporate managers and researchers, software product developers, and engineers in information-intensive companies, such as on-line database and web-service providers.
This book presents a collection of papers on the issue of focus in its broadest sense. While commonly being considered as related to phenomena such as presupposition and anaphora, focusing is much more widely spread, and it is this pervasiveness that this collection addresses. The volume explicitly aims to bring together theoretical, psychological, and descriptive approaches to focus, at the same time maintaining the overall interest in how these notions apply to the larger problem of evolving some formal representation of the semantic aspects of linguistic content. The contributed papers to this volume have been reworked from a selection of original work presented at a conference held in 1994 in Schloss Wolfsbrunnen in Germany.
Originally published in 1997, this book is concerned with human language technology. This technology provides computers with the capability to handle spoken and written language. One major goal is to improve communication between humans and machines. If people can use their own language to access information, working with software applications and controlling machinery, the greatest obstacle for the acceptance of new information technology is overcome. Another important goal is to facilitate communication among people. Machines can help to translate texts or spoken input from one human language to the other. Programs that assist people in writing by checking orthography, grammar and style are constantly improving. This book was sponsored by the Directorate General XIII of the European Union and the Information Science and Engineering Directorate of the National Science Foundation, USA.
CICLing 2004 was the 5th Annual Conference on Intelligent Text Processing and Computational Linguistics; see www.CICLing.org. CICLing conferences are intended to provide a balanced view of the cutting-edge developments in both theoretical foundations of computational linguistics and the practice of natural language text processing with its numerous applications. A feature of CICLing conferences is their wide scope that covers nearly all areas of computational linguistics and all aspects of natural language processing applications. These conferences are a forum for dialogue between the specialists working in the two areas. This year we were honored by the presence of our invited speakers Martin KayofStanfordUniversity, PhilipResnikoftheUniversityofMaryland, Ricardo Baeza-Yates of the University of Chile, and Nick Campbell of the ATR Spoken Language Translation Research Laboratories. They delivered excellent extended lectures and organized vivid discussions. Of129submissionsreceived(74fullpapersand44shortpapers), aftercareful international reviewing 74 papers were selected for presentation (40 full papers and35shortpapers), writtenby176authorsfrom21countries: Korea(37), Spain (34), Japan (22), Mexico (15), China (11), Germany (10), Ireland (10), UK (10), Singapore (6), Canada (3), Czech Rep. (3), France (3), Brazil (2), Sweden (2), Taiwan (2), Turkey (2), USA (2), Chile (1), Romania (1), Thailand (1), and The Netherlands (1); the ?gures in parentheses stand for the number of authors from the corresponding co
The structure and properties of any natural language expression depend on its component sub-expressions - "resources" - and relations among them that are sensitive to basic structural properties of order, grouping, and multiplicity. Resource-sensitivity thus provides a perspective on linguistic structure that is well-defined and universally-applicable. The papers in this collection - by J. van Benthem, P. Jacobson, G. JAger, G-J. Kruijff, G. Morrill, R. Muskens, R. Oehrle, and A. Szabolcsi - examine linguistic resources and resource-sensitivity from a variety of perspectives, including: - Modal aspects of categorial type inference; In particular, the book contains a number of papers treating anaphorically-dependent expressions as functions, whose application to an appropriate argument yields a type and an interpretation directly integratable with the surrounding grammatical structure. To situate this work in a larger setting, the book contains two appendices: - an introductory guide to resource-sensivity;
This book constitutes the refereed proceedings of the 4th International Conference on Computational Linguistics and Intelligent Text Processing, CICLing 2003, held in Mexico City, Mexico in February 2003. The 67 revised papers presented together with 4 keynote papers were carefully reviewed and selected from 92 submissions. The papers are organized in topical sections on computational linguistics formalisms; semantics and discourse; syntax and POS tagging; parsing techniques; morphology; word sense disambiguation; dictionary, lexicon, and ontology; corpus and language statistics; machine translation and bilingual corpora; text generation; natural language interfaces; speech processing; information retrieval and information extraction; text categorization and clustering; summarization; and spell-checking.
This book presents recent advances by leading researchers in computational modelling of language acquisition. The contributors, from departments of linguistics, cognitive science, psychology, and computer science, combine powerful computational techniques with real data and in doing so throw new light on the operations of the brain and the mind. They explore the extent to which linguistic structure is innate and/or available in a child's environment, and the degree to which language learning is inductive or deductive. They assess the explanatory power of different models. The book will appeal to all those working in language acquisition.
CICLing 2001 is the second annual Conference on Intelligent text processing and Computational Linguistics (hence the name CICLing), see www.CICLing.org. It is intended to provide a balanced view of the cutting edge developments in both theoretical foundations of computational linguistics and practice of natural language text processing with its numerous applications. A feature of the CICLing conferences is their wide scope that covers nearly all areas of computational linguistics and all aspects of natural language processing applications. The conference is a forum for dialogue between the specialists working in these two areas. This year our invited speakers were Graeme Hirst (U. Toronto, Canada), Sylvain Kahane (U. Paris 7, France), and Ruslan Mitkov (U. Wolverhampton, UK). They delivered excellent extended lectures and organized vivid discussions. A total of 72 submissions were received, all but very few of surprisingly high quality. After careful reviewing, the Program Committee selected for presentation 53 of them, 41 as full papers and 12 as short papers, by 98 authors from 19 countries: Spain (19 authors), Japan (15), USA (12), France, Mexico (9 each), Sweden (6), Canada, China, Germany, Italy, Malaysia, Russia, United Arab Emirates (3 each), Argentina (2), Bulgaria, The Netherlands, Ukraine, UK, and Uruguay (1 each).
This book constitutes the refereed proceedings of the 4th
International Conference on Text, Speech and Dialogue, TSD 2001,
held in Zelezna Ruda, Czech Republic in September 2001.
This book constitutes the refereed proceedings of the scientific
track of the 7th Congress of the Italian Association for Artificial
Intelligence, AI*IA 2001, held in Bari, Italy, in September
2001.
First published a decade ago, "CJKV Information Processing" quickly became the unsurpassed source of information on processing text in Chinese, Japanese, Korean, and Vietnamese. The book has now been thoroughly updated to provide web and application developers with the latest techniques and tools for disseminating information directly to audiences in East Asia. This new second edition reflects the considerable impact that Unicode, XML, OpenType, and newer operating systems such as Windows XP, Vista, Mac OS X, and Linux have had on East Asian text processing in recent years. Written by its original author, Ken Lunde, a Senior Computer Scientist in CJKV Type Development at Adobe Systems, this book will help you: learn about CJKV writing systems and scripts, and their transliteration methods; explore recent trends and developments in character sets and encodings, particularly Unicode; examine the world of typography, specifically how CJKV text is laid out on a page; learn information processing techniques, such as code conversion algorithms and how to apply them using different programming languages; process CJKV text using different platforms, text editors, and word processors; become more informed about CJKV dictionaries, dictionary software, and machine translation software and services; and, manage CJKV content and presentation when publishing in print or for the Web And much more. Internationalizing and localizing applications is paramount in today's global market - especially for audiences in East Asia, the fastest-growing segment of the computing world. "CJKV Information Processing" will help you understand how to develop web and other applications effectively in a field that many find difficult to master.
This 1992 collection takes the exciting step of examining natural language phenomena from the perspective of both computational linguistics and formal semantics. Computational linguistics has until now been primarily concerned with the construction of computational models for handling the complexities of linguistic form, but has not tackled the questions of representing or computing meaning. Formal semantics, on the other hand, has attempted to account for the relations between forms and meanings, without necessarily attending to computational concerns. The book introduces the reader to the two disciplines and considers the prospects for the more unified and comprehensive computational theory of language which might obtain from their amalgamation. Of great interest to those working in the fields of computation, logic, semantics, artificial intelligence and linguistics generally.
This 1992 collection takes the exciting step of examining natural language phenomena from the perspective of both computational linguistics and formal semantics. Computational linguistics has until now been primarily concerned with the construction of computational models for handling the complexities of linguistic form, but has not tackled the questions of representing or computing meaning. Formal semantics, on the other hand, has attempted to account for the relations between forms and meanings, without necessarily attending to computational concerns. The book introduces the reader to the two disciplines and considers the prospects for the more unified and comprehensive computational theory of language which might obtain from their amalgamation. Of great interest to those working in the fields of computation, logic, semantics, artificial intelligence and linguistics generally.
Anaphora is a central topic in syntax, semantics, and pragmatics and to the interface between them. It is the subject of advanced undergraduate and graduate courses in linguistics and computational linguistics. In this book, Yan Huang provides an extensive and accessible overview of the major contemporary issues surrounding anaphora and gives a critical survey of the many and diverse contemporary approaches to it. He also provides by far the fullest cross-linguistic account of anaphora yet published.
One of the most hotly debated phenomena in natural language is that of leftward argument scrambling. This book investigates the properties of Hindi-Urdu scrambling to show that it must be analyzed as uniformly a focality-driven XP-adjunction operation. It proposes a novel theory of binding and coreference that not only derives the coreference effects in scrambled constructions, but has important consequences for the proper formulation of binding, crossover, reconstruction, and representational economy in the minimalist program. The book will be of interest not only to specialists in Hindi-Urdu syntax and/or scrambling, but to all students of generative syntax.
Understanding any communication depends on the listener or reader recognizing that some words refer to what has already been said or written (his, its, he, there, etc.). This mode of reference, anaphora, involves complicated cognitive and syntactic processes, which people usually perform unerringly, but which present formidable problems for the linguist and cognitive scientist trying to explain precisely how comprehension is achieved. Anaphora is thus a central research focus in syntactic and semantic theory, while understanding and modelling its operation in discourse are important targets in computational linguistics and cognitive science. Yan Huang provides an extensive and accessible overview of the major contemporary issues surrounding anaphora and gives a critical survey of the many and diverse contemporary approaches to it. He provides by far the fullest cross-linguistic account yet published: Dr Huang's survey and analysis are based on a rich collection of data drawn from around 450 of the world's languages.
Information extraction (IE) is a new technology enabling relevant content to be extracted from textual information available electronically. IE essentially builds on natural language processing and computational linguistics, but it is also closely related to the well established area of information retrieval and involves learning. In concert with other promising intelligent information processing technologies like data mining, intelligent data analysis, text summarization, and information agents, IE plays a crucial role in dealing with the vast amounts of information accessible electronically, for example from the Internet. The book is based on the Second International School on Information Extraction, SCIE-99, held in Frascati near Rome, Italy in June/July 1999.
A primary problem in the area of natural language processing has been that of semantic analysis. This book aims to look at the semantics of natural languages in context. It presents an approach to the computational processing of English text that combines current theories of knowledge representation and reasoning in Artificial Intelligence with the latest linguistic views of lexical semantics. This results in distinct advantages for relating the semantic analysis of a sentence to its context. A key feature is the clear separation of the lexical entries that represent the domain-specific linguistic information from the semantic interpreter that performs the analysis. The criteria for defining the lexical entries are firmly grounded in current linguistic theories, facilitating integration with existing parsers. This approach has been implemented and tested in Prolog on a domain for physics word problems and full details of the algorithms and code are presented. Semantic Processing for Finite Domains will appeal to postgraduates and researchers in computational linguistics, and to industrial groups specializing in natural language processing.
This book deals with a major problem in the study of language: the problem of reference. The ease with which we refer to things in conversation is deceptive. Upon closer scrutiny, it turns out that we hardly ever tell each other explicitly what object we mean, although we expect our interlocutor to discern it. Amichai Kronfeld provides an answer to two questions associated with this: how do we successfully refer, and how can a computer be programmed to achieve this? Beginning with the major theories of reference, Dr Kronfeld provides a consistent philosophical view which is a synthesis of Frege's and Russell's semantic insights with Grice's and Searle's pragmatic theories. This leads to a set of guiding principles, which are then applied to a computational model of referring. The discussion is made accessible to readers from a number of backgrounds: in particular, students and researchers in the areas of computational linguistics, artificial intelligence and the philosophy of language will want to read this book.
The goal of this book is to integrate the research being carried out in the field of lexical semantics in linguistics with the work on knowledge representation and lexicon design in computational linguistics. Rarely do these two camps meet and discuss the demands and concerns of each other's fields. Therefore, this book is interesting in that it provides a stimulating and unique discussion between the computational perspective of lexical meaning and the concerns of the linguist for the semantic description of lexical items in the context of syntactic descriptions. This book grew out of the papers presented at a workshop held at Brandeis University in April, 1988, funded by the American Association for Artificial Intelligence. The entire workshop as well as the discussion periods accom panying each talk were recorded. Once complete copies of each paper were available, they were distributed to participants, who were asked to provide written comments on the texts for review purposes. VII JAMES PUSTEJOVSKY 1. INTRODUCTION There is currently a growing interest in the content of lexical entries from a theoretical perspective as well as a growing need to understand the organization of the lexicon from a computational view. This volume attempts to define the directions that need to be taken in order to achieve the goal of a coherent theory of lexical organization."
From tech giants to plucky startups, the world is full of companies boasting that they are on their way to replacing human interpreters, but are they right? Interpreters vs Machines offers a solid introduction to recent theory and research on human and machine interpreting, and then invites the reader to explore the future of interpreting. With a foreword by Dr Henry Liu, the 13th International Federation of Translators (FIT) President, and written by consultant interpreter and researcher Jonathan Downie, this book offers a unique combination of research and practical insight into the field of interpreting. Written in an innovative, accessible style with humorous touches and real-life case studies, this book is structured around the metaphor of playing and winning a computer game. It takes interpreters of all experience levels on a journey to better understand their own work, learn how computers attempt to interpret and explore possible futures for human interpreters. With five levels and split into 14 chapters, Interpreters vs Machines is key reading for all professional interpreters as well as students and researchers of Interpreting and Translation Studies, and those with an interest in machine interpreting.
This book is an advanced introduction to semantics that presents this crucial component of human language through the lens of the 'Meaning-Text' theory - an approach that treats linguistic knowledge as a huge inventory of correspondences between thought and speech. Formally, semantics is viewed as an organized set of rules that connect a representation of meaning (Semantic Representation) to a representation of the sentence (Deep-Syntactic Representation). The approach is particularly interesting for computer assisted language learning, natural language processing and computational lexicography, as our linguistic rules easily lend themselves to formalization and computer applications. The model combines abstract theoretical constructions with numerous linguistic descriptions, as well as multiple practice exercises that provide a solid hands-on approach to learning how to describe natural language semantics.
A comprehensive reference book for detailed explanations for every algorithm and techniques related to the transformers. 60+ transformer architectures covered in a comprehensive manner. A book for understanding how to apply the transformer techniques in speech, text, time series, and computer vision. Practical tips and tricks for each architecture and how to use it in the real world. Hands-on case studies and code snippets for theory and practical real-world analysis using the tools and libraries, all ready to run in Google Colab. |
You may like...
A Mafia Romance - Never Been Caught (A…
Michelle Love, Ivy Wonder
Hardcover
R857
Discovery Miles 8 570
Advanced Logic Synthesis
Andre Inacio Reis, Rolf Drechsler
Hardcover
|