![]() |
![]() |
Your cart is empty |
||
Books > Language & Literature > Language & linguistics > Computational linguistics
This book explains how to build Natural Language Generation (NLG) systems - computer software systems which use techniques from artificial intelligence and computational linguistics to automatically generate understandable texts in English or other human languages, either in isolation or as part of multimedia documents, Web pages, and speech output systems. Typically starting from some non-linguistic representation of information as input, NLG systems use knowledge about language and the application domain to automatically produce documents, reports, explanations, help messages, and other kinds of texts. The book covers the algorithms and representations needed to perform the core tasks of document planning, microplanning, and surface realization, using a case study to show how these components fit together. It also discusses engineering issues such as system architecture, requirements analysis, and the integration of text generation into multimedia and speech output systems.
This is a collection of new papers by leading researchers on natural language parsing. In the past, the problem of how people parse the sentences they hear - determine the identity of the words in these sentences and group these words into larger units - has been addressed in very different ways by experimental psychologists, by theoretical linguists, and by researchers in artificial intelligence, with little apparent relationship among the solutions proposed by each group. However, because of important advances in all these disciplines, research on parsing in each of these fields now seems to have something significant to contribute to the others, as this volume demonstrates. The volume includes some papers applying the results of experimental psychological studies of parsing to linguistic theory, others which present computational models of parsing, and a mathematical linguistics paper on tree-adjoining grammars and parsing.
People often mean more than they say. Grammar on its own is typically insufficient for determining the full meaning of an utterance; the assumption that the discourse is coherent or 'makes sense' has an important role to play in determining meaning as well. Logics of Conversation presents a dynamic semantic framework called Segmented Discourse Representation Theory, or SDRT, where this interaction between discourse coherence and discourse interpretation is explored in a logically precise manner. Combining ideas from dynamic semantics, commonsense reasoning and speech act theory, SDRT uses its analysis of rhetorical relations to capture intuitively compelling implicatures. It provides a computable method for constructing these logical forms and is one of the most formally precise and linguistically grounded accounts of discourse interpretation currently available. The book will be of interest to researchers and students in linguistics and in philosophy of language.
Originally published in 1997, this book is concerned with human language technology. This technology provides computers with the capability to handle spoken and written language. One major goal is to improve communication between humans and machines. If people can use their own language to access information, working with software applications and controlling machinery, the greatest obstacle for the acceptance of new information technology is overcome. Another important goal is to facilitate communication among people. Machines can help to translate texts or spoken input from one human language to the other. Programs that assist people in writing by checking orthography, grammar and style are constantly improving. This book was sponsored by the Directorate General XIII of the European Union and the Information Science and Engineering Directorate of the National Science Foundation, USA.
This book provides a computational re-evaluation of the genealogical relations between the early Germanic families and of their diversification from their most recent common ancestor, Proto-Germanic. It also proposes a novel computational approach to the problem of linguistic diversification more broadly, using agent-based simulation of speech communities over time. This new method is presented alongside more traditional phylogenetic inference, and the respective results are compared and evaluated. Frederik Hartmann demonstrates that the traditional and novel methods each capture different aspects of this highly complex real-world process; crucially, the new computational approach proposed here offers a new way of investigating the wave-like properties of language relatedness that were previously less accessible. As well as validating the findings of earlier research, the results of this study also generate new insights and shed light on much-debated issues in the field. The conclusion is that the break-up of Germanic should be understood as a gradual disintegration process in which tree-like branching effects are rare.
This 1992 collection takes the exciting step of examining natural language phenomena from the perspective of both computational linguistics and formal semantics. Computational linguistics has until now been primarily concerned with the construction of computational models for handling the complexities of linguistic form, but has not tackled the questions of representing or computing meaning. Formal semantics, on the other hand, has attempted to account for the relations between forms and meanings, without necessarily attending to computational concerns. The book introduces the reader to the two disciplines and considers the prospects for the more unified and comprehensive computational theory of language which might obtain from their amalgamation. Of great interest to those working in the fields of computation, logic, semantics, artificial intelligence and linguistics generally.
Semantic interpretation and the resolution of ambiguity presents an important advance in computer understanding of natural language. While parsing techniques have been greatly improved in recent years, the approach to semantics has generally improved in recent years, the approach to semantics has generally been ad hoc and had little theoretical basis. Graeme Hirst offers a new, theoretically motivated foundation for conceptual analysis by computer, and shows how this framework facilitates the resolution of lexical and syntactic ambiguities. His approach is interdisciplinary, drawing on research in computational linguistics, artificial intelligence, montague semantics, and cognitive psychology.
A primary problem in the area of natural language processing has been that of semantic analysis. This book aims to look at the semantics of natural languages in context. It presents an approach to the computational processing of English text that combines current theories of knowledge representation and reasoning in Artificial Intelligence with the latest linguistic views of lexical semantics. This results in distinct advantages for relating the semantic analysis of a sentence to its context. A key feature is the clear separation of the lexical entries that represent the domain-specific linguistic information from the semantic interpreter that performs the analysis. The criteria for defining the lexical entries are firmly grounded in current linguistic theories, facilitating integration with existing parsers. This approach has been implemented and tested in Prolog on a domain for physics word problems and full details of the algorithms and code are presented. Semantic Processing for Finite Domains will appeal to postgraduates and researchers in computational linguistics, and to industrial groups specializing in natural language processing.
This book deals with a major problem in the study of language: the problem of reference. The ease with which we refer to things in conversation is deceptive. Upon closer scrutiny, it turns out that we hardly ever tell each other explicitly what object we mean, although we expect our interlocutor to discern it. Amichai Kronfeld provides an answer to two questions associated with this: how do we successfully refer, and how can a computer be programmed to achieve this? Beginning with the major theories of reference, Dr Kronfeld provides a consistent philosophical view which is a synthesis of Frege's and Russell's semantic insights with Grice's and Searle's pragmatic theories. This leads to a set of guiding principles, which are then applied to a computational model of referring. The discussion is made accessible to readers from a number of backgrounds: in particular, students and researchers in the areas of computational linguistics, artificial intelligence and the philosophy of language will want to read this book.
The idea that the expression of radical beliefs is a predictor to future acts of political violence has been a central tenet of counter-extremism over the last two decades. Not only has this imposed a duty upon doctors, lecturers and teachers to inform on the radical beliefs of their patients and students but, as this book argues, it is also a fundamentally flawed concept. Informed by his own experience with the UK's Prevent programme while teaching in a Muslim community, Rob Faure Walker explores the linguistic emergence of 'extremism' in political discourse and the potentially damaging generative effect of this language. Taking a new approach which combines critical discourse analysis with critical realism, this book shows how the fear of being labelled as an 'extremist' has resulted in counter-terrorism strategies which actually undermine moderating mechanisms in a democracy. Analysing the generative mechanisms by which the language of counter-extremism might actually promote violence, Faure Walker explains how understanding the potentially oppressive properties of language can help us transcend them. The result is an imminent critique of the most pernicious aspects of the global War on Terror, those that are embedded in our everyday language and political discourse. Drawing on the author's own successful lobbying activities against counter-extremism, this book presents a model for how discourse analysis and critical realism can and should engage with the political and how this will affect meaningful change.
This book is an advanced introduction to semantics that presents this crucial component of human language through the lens of the 'Meaning-Text' theory - an approach that treats linguistic knowledge as a huge inventory of correspondences between thought and speech. Formally, semantics is viewed as an organized set of rules that connect a representation of meaning (Semantic Representation) to a representation of the sentence (Deep-Syntactic Representation). The approach is particularly interesting for computer assisted language learning, natural language processing and computational lexicography, as our linguistic rules easily lend themselves to formalization and computer applications. The model combines abstract theoretical constructions with numerous linguistic descriptions, as well as multiple practice exercises that provide a solid hands-on approach to learning how to describe natural language semantics.
In this brief, the authors discuss recently explored spectral (sub-segmental and pitch synchronous) and prosodic (global and local features at word and syllable levels in different parts of the utterance) features for discerning emotions in a robust manner. The authors also delve into the complementary evidences obtained from excitation source, vocal tract system and prosodic features for the purpose of enhancing emotion recognition performance. Features based on speaking rate characteristics are explored with the help of multi-stage and hybrid models for further improving emotion recognition performance. Proposed spectral and prosodic features are evaluated on real life emotional speech corpus.
In everyday communication, Europe's citizens, business partners and politicians are inevitably confronted with language barriers. Language technology has the potential to overcome these barriers and to provide innovative interfaces to technologies and knowledge. This document presents a Strategic Research Agenda for Multilingual Europe 2020. The agenda was prepared by META-NET, a European Network of Excellence. META-NET consists of 60 research centres in 34 countries, who cooperate with stakeholders from economy, government agencies, research organisations, non-governmental organisations, language communities and European universities. META-NET's vision is high-quality language technology for all European languages. "The research carried out in the area of language technology is of utmost importance for the consolidation of Portuguese as a language of global communication in the information society." - Dr. Pedro Passos Coelho (Prime-Minister of Portugal) "It is imperative that language technologies for Slovene are developed systematically if we want Slovene to flourish also in the future digital world." - Dr. Danilo Turk (President of the Republic of Slovenia) "For such small languages like Latvian keeping up with the ever increasing pace of time and technological development is crucial. The only way to ensure future existence of our language is to provide its users with equal opportunities as the users of larger languages enjoy. Therefore being on the forefront of modern technologies is our opportunity." - Valdis Dombrovskis (Prime Minister of Latvia) "Europe's inherent multilingualism and our scientific expertise are the perfect prerequisites for significantly advancing the challenge that language technology poses. META-NET opens up new opportunities for the development of ubiquitous multilingual technologies." - Prof. Dr. Annette Schavan (German Minister of Education and Research)
This book explores the interaction between corpus stylistics and translation studies. It shows how corpus methods can be used to compare literary texts to their translations, through the analysis of Joseph Conrad's Heart of Darkness and four of its Italian translations. The comparison focuses on stylistic features related to the major themes of Heart of Darkness. By combining quantitative and qualitative techniques, Mastropierro discusses how alterations to the original's stylistic features can affect the interpretation of the themes in translation. The discussion illuminates the manipulative effects that translating can have on the reception of a text, showing how textual alterations can trigger different readings. This book advances the multidisciplinary dialogue between corpus linguistics and translation studies and is a valuable resource for students and researchers interested in the application of corpus approaches to stylistics and translation.
Die Entwicklung und Verbreitung von Systemen fur maschinelles UEbersetzen bewirkt massive Transformationsprozesse in der Sprachdienstleistungsbranche. Die 'Maschinisierung' von Translation fuhrt nicht nur zu Umwalzungen innerhalb des UEbersetzungsmarktes, sondern stellt uns auch vor die grundlegende Frage: Was ist 'UEbersetzen', wenn eine Maschine menschliche Sprache ubersetzt? Diese Arbeit widmet sich diesem Problem aus der Perspektive der Translationswissenschaft und der Techniksoziologie. Im Fokus stehen Translationskonzepte in der Computerlinguistik, die aus einer Wechselwirkung zwischen sozialer Konstruktion und technischen Gegebenheiten resultieren. Der UEbersetzungsbegriff von Computerlinguist:innen orientiert sich an der Mechanik der Maschine, wodurch ein Spannungsverhaltnis mit den Paradigmen der Humantranslation entsteht.
This book is open access and available on www.bloomsburycollections.com. It is funded by Knowledge Unlatched. Corpus linguistics has much to offer history, being as both disciplines engage so heavily in analysis of large amounts of textual material. This book demonstrates the opportunities for exploring corpus linguistics as a method in historiography and the humanities and social sciences more generally. Focussing on the topic of prostitution in 17th-century England, it shows how corpus methods can assist in social research, and can be used to deepen our understanding and comprehension. McEnery and Baker draw principally on two sources - the newsbook Mercurius Fumigosis and the Early English Books Online Corpus. This scholarship on prostitution and the sex trade offers insight into the social position of women in history.
In diesem Open-Access-Buch wird mithilfe eines grossangelegten Online-Experiments untersucht, wie sich die Anzeige von Zitationen oder Downloads auf die Relevanzbewertung in akademischen Suchsystemen auswirkt. Bei der Suche nach Informationen verwenden Menschen diverse Kriterien, anhand derer sie die Relevanz der Suchergebnisse bewerten. In diesem Buch wird erstmals eine systematische UEbersicht uber die Einflusse im Prozess der Relevanzbewertung von Suchergebnissen in akademischen Suchsystemen aufgezeigt. Zudem wird ein anspruchsvolles und komplexes Methodenframework zur experimentellen Untersuchung von Relevanzkriterien vorgestellt. Dieses eignet sich fur die weitergehende Erforschung von Relevanzkriterien im informationswissenschaftlichen Bereich.
When we speak, we configure the vocal tract which shapes the visible motions of the face and the patterning of the audible speech acoustics. Similarly, we use these visible and audible behaviors to perceive speech. This book showcases a broad range of research investigating how these two types of signals are used in spoken communication, how they interact, and how they can be used to enhance the realistic synthesis and recognition of audible and visible speech. The volume begins by addressing two important questions about human audiovisual performance: how auditory and visual signals combine to access the mental lexicon and where in the brain this and related processes take place. It then turns to the production and perception of multimodal speech and how structures are coordinated within and across the two modalities. Finally, the book presents overviews and recent developments in machine-based speech recognition and synthesis of AV speech.
Language and Computers introduces students to the fundamentals of how computers are used to represent, process, and organize textual and spoken information. Concepts are grounded in real-world examples familiar to students experiences of using language and computers in everyday life. * A real-world introduction to the fundamentals of how computers process language, written specifically for the undergraduate audience, introducing key concepts from computational linguistics. * Offers a comprehensive explanation of the problems computers face in handling natural language * Covers a broad spectrum of language-related applications and issues, including major computer applications involving natural language and the social and ethical implications of these new developments * The book focuses on real-world examples with which students can identify, using these to explore the technology and how it works * Features under-the-hood sections that give greater detail on selected advanced topics, rendering the book appropriate for more advanced courses, or for independent study by the motivated reader.
Experimental syntax is an area that is rapidly growing as linguistic research becomes increasingly focused on replicable language data, in both fieldwork and laboratory environments. The first of its kind, this handbook provides an in-depth overview of current issues and trends in this field, with contributions from leading international scholars. It pays special attention to sentence acceptability experiments, outlining current best practices in conducting tests, and pointing out promising new avenues for future research. Separate sections review research results from the past 20 years, covering specific syntactic phenomena and language types. The handbook also outlines other common psycholinguistic and neurolinguistic methods for studying syntax, comparing and contrasting them with acceptability experiments, and giving useful perspectives on the interplay between theoretical and experimental linguistics. Providing an up-to-date reference on this exciting field, it is essential reading for students and researchers in linguistics interested in using experimental methods to conduct syntactic research.
This book is about a new approach in the field of computational linguistics related to the idea of constructing n-grams in non-linear manner, while the traditional approach consists in using the data from the surface structure of texts, i.e., the linear structure.In this book, we propose and systematize the concept of syntactic n-grams, which allows using syntactic information within the automatic text processing methods related to classification or clustering. It is a very interesting example of application of linguistic information in the automatic (computational) methods. Roughly speaking, the suggestion is to follow syntactic trees and construct n-grams based on paths in these trees. There are several types of non-linear n-grams; future work should determine, which types of n-grams are more useful in which natural language processing (NLP) tasks. This book is intended for specialists in the field of computational linguistics. However, we made an effort to explain in a clear manner how to use n-grams; we provide a large number of examples, and therefore we believe that the book is also useful for graduate students who already have some previous background in the field.
This white paper is part of a series that promotes knowledge about language technology and its potential. It addresses educators, journalists, politicians, language communities and others. The availability and use of language technology in Europe varies between languages. Consequently, the actions that are required to further support research and development of language technologies also differ for each language. The required actions depend on many factors, such as the complexity of a given language and the size of its community. META-NET, a Network of Excellence funded by the European Commission, has conducted an analysis of current language resources and technologies. This analysis focused on the 23 official European languages as well as other important national and regional languages in Europe. The results of this analysis suggest that there are many significant research gaps for each language. A more detailed expert analysis and assessment of the current situation will help maximise the impact of additional research and minimize any risks. META-NET consists of 54 research centres from 33 countries that are working with stakeholders from commercial businesses, government agencies, industry, research organisations, software companies, technology providers and European universities. Together, they are creating a common technology vision while developing a strategic research agenda that shows how language technology applications can address any research gaps by 2020.
Specifically designed for linguists, this book provides an introduction to programming using Python for those with little to no experience of coding. Python is one of the most popular and widely-used programming languages as it's also available for free and runs on any operating system. All examples in the text involve language data and can be adapted or used directly for language research. The text focuses on key language-related issues: searching, text manipulation, text encoding and internet data, providing an excellent resource for language research. More experienced users of Python will also benefit from the advanced chapters on graphical user interfaces and functional programming. |
![]() ![]() You may like...
Biomedical Diagnostics and Clinical…
Manuela Pereira, Mario Freire
Hardcover
R7,022
Discovery Miles 70 220
Fixed-Time Cooperative Control of…
Zongyu Zuo, Qinglong Han, …
Hardcover
R3,538
Discovery Miles 35 380
Finite Volumes for Complex Applications…
Clement Cances, Pascal Omnes
Hardcover
R5,412
Discovery Miles 54 120
|