![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Language & Literature > Language & linguistics > Computational linguistics
Cross-Disciplinary Advances in Applied Natural Language Processing: Issues and Approaches defines the role of ANLP within NLP, and alongside other disciplines such as linguistics, computer science, and cognitive science. The description also includes the categorization of current ANLP research, and examples of current research in ANLP. This book is a useful reference for teachers, students, and materials developers in fields spanning linguistics, computer science, and cognitive science.
Tense and aspect are means by which language refers to time-how an event takes place in the past, present, or future. They play a key role in understanding the grammar and structure of all languages, and interest in them reaches across linguistics. The Oxford Handbook of Tense and Aspect is a comprehensive, authoritative, and accessible guide to the topics and theories that currently form the front line of research into tense, aspect, and related areas. The volume contains 36 chapters, divided into 6 sections, written by internationally known experts in theoretical linguistics.
This book presents the consolidated acoustic data for all phones in Standard Colloquial Bengali (SCB), commonly known as Bangla, a Bengali language used by 350 million people in India, Bangladesh, and the Bengali diaspora. The book analyzes the real speech of selected native speakers of the Bangla dialect to ensure that a proper acoustical database is available for the development of speech technologies. The acoustic data presented consists of averages and their normal spread, represented by the standard deviations of necessary acoustic parameters including e.g. formant information for multiple native speakers of both sexes. The study employs two important speech technologies:(1) text to speech synthesis (TTS) and (2) automatic speech recognition (ASR). The procedures, particularly those related to the use of technologies, are described in sufficient detail to enable researchers to use them to create technical acoustic databases for any other Indian dialect. The book offers a unique resource for scientists and industrial practitioners who are interested in the acoustic analysis and processing of Indian dialects to develop similar dialect databases of their own.
Collaboratively Constructed Language Resources (CCLRs) such as
Wikipedia, Wiktionary, Linked Open Data, and various resources
developed using crowdsourcing techniques such as Games with a
Purpose and Mechanical Turk have substantially contributed to the
research in natural language processing (NLP). Various NLP tasks
utilize such resources to substitute for or supplement conventional
lexical semantic resources and linguistically annotated corpora.
These resources also provide an extensive body of texts from which
valuable knowledge is mined. There are an increasing number of
community efforts to link and maintain multiple linguistic
resources.
"Corpora and Language Education" critically examines key concepts and issues in corpus linguistics, with a particular focus on the expanding interdisciplinary nature of the field and the role that written and spoken corpora now play in the fields of professional communication, teacher education, translation studies, lexicography, literature, critical discourse analysis and forensic linguistics. The book also presents a series of corpus-based case studies illustrating central themes and best practices in the field.
This book brings together scientists, researchers, practitioners, and students from academia and industry to present recent and ongoing research activities concerning the latest advances, techniques, and applications of natural language processing systems, and to promote the exchange of new ideas and lessons learned. Taken together, the chapters of this book provide a collection of high-quality research works that address broad challenges in both theoretical and applied aspects of intelligent natural language processing. The book presents the state-of-the-art in research on natural language processing, computational linguistics, applied Arabic linguistics and related areas. New trends in natural language processing systems are rapidly emerging - and finding application in various domains including education, travel and tourism, and healthcare, among others. Many issues encountered during the development of these applications can be resolved by incorporating language technology solutions. The topics covered by the book include: Character and Speech Recognition; Morphological, Syntactic, and Semantic Processing; Information Extraction; Information Retrieval and Question Answering; Text Classification and Text Mining; Text Summarization; Sentiment Analysis; Machine Translation Building and Evaluating Linguistic Resources; and Intelligent Language Tutoring Systems.
Recent advances in the fields of knowledge representation, reasoning and human-computer interaction have paved the way for a novel approach to treating and handling context. The field of research presented in this book addresses the problem of contextual computing in artificial intelligence based on the state of the art in knowledge representation and human-computer interaction. The author puts forward a knowledge-based approach for employing high-level context in order to solve some persistent and challenging problems in the chosen showcase domain of natural language understanding. Specifically, the problems addressed concern the handling of noise due to speech recognition errors, semantic ambiguities, and the notorious problem of underspecification. Consequently the book examines the individual contributions of contextual composing for different types of context. Therefore, contextual information stemming from the domain at hand, prior discourse, and the specific user and real world situation are considered and integrated in a formal model that is applied and evaluated employing different multimodal mobile dialog systems. This book is intended to meet the needs of readers from at least three fields - AI and computer science; computational linguistics; and natural language processing - as well as some computationally oriented linguists, making it a valuable resource for scientists, researchers, lecturers, language processing practitioners and professionals as well as postgraduates and some undergraduates in the aforementioned fields. "The book addresses a problem of great and increasing technical and practical importance - the role of context in natural language processing (NLP). It considers the role of context in three important tasks: Automatic Speech Recognition, Semantic Interpretation, and Pragmatic Interpretation. Overall, the book represents a novel and insightful investigation into the potential of contextual information processing in NLP." Jerome A Feldman, Professor of Electrical Engineering and Computer Science, UC Berkeley, USA http://dm.tzi.de/research/contextual-computing/
The contributions to this volume offer a broad range of novel insights about data-based or data-driven approaches to the study of both structure and function of language, reflecting the increasing shift towards corpus-based methods of analysis in a wide range of areas in linguistics. Corpora can be used as models of human linguistic experience, and the contributors demonstrate that there is ample scope for integrating such models into the descriptions of discourse, grammar, and meaning. Continually improving technological development facilitates the design of larger and more comprehensive corpora documenting language use in a multitude of genres, styles and modes, even starting to include visual aspects. Software to investigate these data also becomes increasingly powerful and more refined. The sixteen original articles in this volume cover substantial ground on both the theoretical as well as applied levels. Having such data and software resources at their disposal, the contributing researchers rethink the long discussed interplay between language system and use from various angles, considering socio-cultural and cognitive involvement and representation, with synchronic as well as diachronic perspectives in view. These theories and quantitative / qualitative methods are applied to a range of topics from language acquisition and teaching to literature and politics. All of the authors in this volume reveal the profound and leading impact that Mike Stubbs' work has continued to contribute to the field of corpus-based description of language structure, use and function.
The contributions in this volume focus on the Bayesian interpretation of natural languages, which is widely used in areas of artificial intelligence, cognitive science, and computational linguistics. This is the first volume to take up topics in Bayesian Natural Language Interpretation and make proposals based on information theory, probability theory, and related fields. The methodologies offered here extend to the target semantic and pragmatic analyses of computational natural language interpretation. Bayesian approaches to natural language semantics and pragmatics are based on methods from signal processing and the causal Bayesian models pioneered by especially Pearl. In signal processing, the Bayesian method finds the most probable interpretation by finding the one that maximizes the product of the prior probability and the likelihood of the interpretation. It thus stresses the importance of a production model for interpretation as in Grice's contributions to pragmatics or in interpretation by abduction.
This book focuses on information literacy for the younger generation of learners and library readers. It is divided into four sections: 1. Information Literacy for Life; 2. Searching Strategies, Disciplines and Special Topics; 3. Information Literacy Tools for Evaluating and Utilizing Resources; 4. Assessment of Learning Outcomes. Written by librarians with wide experience in research and services, and a strong academic background in disciplines such as the humanities, social sciences, information technology, and library science, this valuable reference resource combines both theory and practice. In today's ever-changing era of information, it offers students of library and information studies insights into information literacy as well as learning tips they can use for life.
This book provides a gradual introduction to the naming game, starting from the minimal naming game, where the agents have infinite memories (Chapter 2), before moving on to various new and advanced settings: the naming game with agents possessing finite-sized memories (Chapter 3); the naming game with group discussions (Chapter 4); the naming game with learning errors in communications (Chapter 5) ; the naming game on multi-community networks (Chapter 6) ; the naming game with multiple words or sentences (Chapter 7) ; and the naming game with multiple languages (Chapter 8). Presenting the authors' own research findings and developments, the book provides a solid foundation for future advances. This self-study resource is intended for researchers, practitioners, graduate and undergraduate students in the fields of computer science, network science, linguistics, data engineering, statistical physics, social science and applied mathematics.
There is increasing interaction among communities with multiple languages, thus we need services that can effectively support multilingual communication. The Language Grid is an initiative to build an infrastructure that allows end users to create composite language services for intercultural collaboration. The aim is to support communities to create customized multilingual environments by using language services to overcome local language barriers. The stakeholders of the Language Grid are the language resource providers, the language service users, and the language grid operators who coordinate the former. This book includes 18 chapters in six parts that summarize various research results and associated development activities on the Language Grid. The chapters in Part I describe the framework of the Language Grid, i.e., service-oriented collective intelligence, used to bridge providers, users and operators. Two kinds of software are introduced, the service grid server software and the Language Grid Toolbox, and code for both is available via open source licenses. Part II describes technologies for service workflows that compose atomic language services. Part III reports on research work and activities relating to sharing and using language services. Part IV describes various applications of language services as applicable to intercultural collaboration. Part V contains reports on applying the Language Grid for translation activities, including localization of industrial documents and Wikipedia articles. Finally, Part VI illustrates how the Language Grid can be connected to other service grids, such as DFKI's Heart of Gold and smart classroom services in Tsinghua University in Beijing. The book will be valuable for researchers in artificial intelligence, natural language processing, services computing and human--computer interaction, particularly those who are interested in bridging technologies and user communities. "
The accurate determination of the speech spectrum, particularly for short frames, is commonly pursued in diverse areas including speech processing, recognition, and acoustic phonetics. With this book the author makes the subject of spectrum analysis understandable to a wide audience, including those with a solid background in general signal processing and those without such background. In keeping with these goals, this is not a book that replaces or attempts to cover the material found in a general signal processing textbook. Some essential signal processing concepts are presented in the first chapter, but even there the concepts are presented in a generally understandable fashion as far as is possible. Throughout the book, the focus is on applications to speech analysis; mathematical theory is provided for completeness, but these developments are set off in boxes for the benefit of those readers with sufficient background. Other readers may proceed through the main text, where the key results and applications will be presented in general heuristic terms, and illustrated with software routines and practical "show-and-tell" discussions of the results. At some points, the book refers to and uses the implementations in the Praat speech analysis software package, which has the advantages that it is used by many scientists around the world, and it is free and open source software. At other points, special software routines have been developed and made available to complement the book, and these are provided in the Matlab programming language. If the reader has the basic Matlab package, he/she will be able to immediately implement the programs in that platform---no extra "toolboxes" are required.
This book is written for both linguists and computer scientists working in the field of artificial intelligence as well as to anyone interested in intelligent text processing. Lexical function is a concept that formalizes semantic and syntactic relations between lexical units. Collocational relation is a type of institutionalized lexical relations which holds between the base and its partner in a collocation. Knowledge of collocation is important for natural language processing because collocation comprises the restrictions on how words can be used together. The book shows how collocations can be annotated with lexical functions in a computer readable dictionary - allowing their precise semantic analysis in texts and their effective use in natural language applications including parsers, high quality machine translation, periphrasis system and computer-aided learning of lexica. The books shows how to extract collocations from corpora and annotate them with lexical functions automatically. To train algorithms, the authors created a dictionary of lexical functions containing more than 900 Spanish disambiguated and annotated examples which is a part of this book. The obtained results show that machine learning is feasible to achieve the task of automatic detection of lexical functions.
The volume "Genres on the Web" has been designed for a wide audience, from the expert to the novice. It is a required book for scholars, researchers and students who want to become acquainted with the latest theoretical, empirical and computational advances in the expanding field of web genre research. The study of web genre is an overarching and interdisciplinary novel area of research that spans from corpus linguistics, computational linguistics, NLP, and text-technology, to web mining, webometrics, social network analysis and information studies. This book gives readers a thorough grounding in the latest research on web genres and emerging document types. The book covers a wide range of web-genre focused subjects, such
as: One of the driving forces behind genre research is the idea of a genre-sensitive information system, which incorporates genre cues complementing the current keyword-based search and retrieval applications."
area and in applications to linguistics, formal epistemology, and the study of norms. The second contains papers on non-classical and many-valued logics, with an eye on applications in computer science and through it to engineering. The third concerns the logic of belief management, whichis likewise closely connected with recent work in computer science but also links directly with epistemology, the philosophy of science, the study of legal and other normative systems, and cognitive science. The grouping is of course rough, for there are contributions to the volume that lie astride a boundary; at least one of them is relevant, from a very abstract perspective, to all three areas. We say a few words about each of the individual chapters, to relate them to each other and the general outlook of the volume. Modal Logics The ?rst bundle of papers in this volume contains contribution to modal logic. Three of them examine general problems that arise for all kinds of modal logics. The ?rst paper is essentially semantical in its approach, the second proof-theoretic, the third semantical again: Commutativity of quanti?ers in varying-domain Kripke models, by R. Goldblatt and I. Hodkinson, investigates the possibility of com- tation (i.e. reversing the order) for quanti?ers in ?rst-order modal logics interpreted over relational models with varying domains. The authors study a possible-worlds style structural model theory that does not v- idate commutation, but satis?es all the axioms originally presented by Kripke for his familiar semantics for ?rst-order modal logic."
Mathematical Linguistics introduces the mathematical foundations of linguistics to computer scientists, engineers, and mathematicians interested in natural language processing. The book presents linguistics as a cumulative body of knowledge from the ground up: no prior knowledge of linguistics is assumed. Previous textbooks in this area concentrate on syntax and semantics - this comprehensive volume covers an extremely rich array of topics also including phonology and morphology, probabilistic approaches, complexity, learnability, and the analysis of speech and handwriting. As the first textbook of its kind, this book is useful for those in information science (information retrieval and extraction, search engines) and in natural language technologies (speech recognition, optical character recognition, HCI). Exercises suitable for the advanced reader are included, as well as suggestions for further reading and an extensive bibliography.
One of the most hotly debated phenomena in natural language is that of leftward argument scrambling. This book investigates the properties of Hindi-Urdu scrambling to show that it must be analyzed as uniformly a focality-driven XP-adjunction operation. It proposes a novel theory of binding and coreference that not only derives the coreference effects in scrambled constructions, but has important consequences for the proper formulation of binding, crossover, reconstruction, and representational economy in the minimalist program. The book will be of interest not only to specialists in Hindi-Urdu syntax and/or scrambling, but to all students of generative syntax.
Research in Natural Language Processing (NLP) has rapidly advanced in recent years, resulting in exciting algorithms for sophisticated processing of text and speech in various languages. Much of this work focuses on English; in this book we address another group of interesting and challenging languages for NLP research: the Semitic languages. The Semitic group of languages includes Arabic (206 million native speakers), Amharic (27 million), Hebrew (7 million), Tigrinya (6.7 million), Syriac (1 million) and Maltese (419 thousand). Semitic languages exhibit unique morphological processes, challenging syntactic constructions and various other phenomena that are less prevalent in other natural languages. These challenges call for unique solutions, many of which are described in this book. The 13 chapters presented in this book bring together leading scientists from several universities and research institutes worldwide. While this book devotes some attention to cutting-edge algorithms and techniques, its primary purpose is a thorough explication of best practices in the field. Furthermore, every chapter describes how the techniques discussed apply to Semitic languages. The book covers both statistical approaches to NLP, which are dominant across various applications nowadays and the more traditional, rule-based approaches, that were proven useful for several other application domains. We hope that this book will provide a "one-stop-shop'' for all the requisite background and practical advice when building NLP applications for Semitic languages.
This book provides an in-depth description of the framework of inductive dependency parsing, a methodology for robust and efficient syntactic analysis of unrestricted natural language text. This methodology is based on two essential components: dependency-based syntactic representations and a data-driven approach to syntactic parsing. More precisely, it is based on a deterministic parsing algorithm in combination with inductive machine learning to predict the next parser action. The book includes a theoretical analysis of all central models and algorithms, as well as a thorough empirical evaluation of memory-based dependency parsing, using data from Swedish and English. Offering the reader a one-stop reference to dependency-based parsing of natural language, it is intended for researchers and system developers in the language technology field, and is also suited for graduate or advanced undergraduate education.
This book celebrates the work of Yorick Wilks in the form of a selection of his papers which are intended to reflect the range and depth of his work. The volume accompanies a Festschrift which celebrates his contribution to the fields of Computational Linguistics and Artificial Intelligence. The selected papers reflect Yorick 's contribution to both practical and theoretical aspects of automatic language processing.
Academic vocabulary is in fashion, as witnessed by the increasing number of books published on the topic. In the first part of this book , Magali Paquot scrutinizes the concept of 'academic vocabulary' and proposes a corpus-driven procedure based on the criteria of keyness, range and evenness of distribution to select academic words that could be part of a common-core academic vocabulary syllabus. In the second part, the author offers a thorough analysis of academic vocabulary in the International Corpus of Learner English (ICLE) and describes the factors that account for learners' difficulties in academic writing. She then focuses on the role of corpora, and more particularly, learner corpora, in EAP material design. It is the first monograph in which Granger's (1996) Contrastive Interlanguage Analysis is used to compare 10 ICLE learner sub-corpora, in order to distinguish between linguistic features that are shared by learners from a wide range of mother tongue backgrounds and unique features that may be transfer-related.
The structure and properties of any natural language expression depend on its component sub-expressions - "resources" - and relations among them that are sensitive to basic structural properties of order, grouping, and multiplicity. Resource-sensitivity thus provides a perspective on linguistic structure that is well-defined and universally-applicable. The papers in this collection - by J. van Benthem, P. Jacobson, G. JAger, G-J. Kruijff, G. Morrill, R. Muskens, R. Oehrle, and A. Szabolcsi - examine linguistic resources and resource-sensitivity from a variety of perspectives, including: - Modal aspects of categorial type inference; In particular, the book contains a number of papers treating anaphorically-dependent expressions as functions, whose application to an appropriate argument yields a type and an interpretation directly integratable with the surrounding grammatical structure. To situate this work in a larger setting, the book contains two appendices: - an introductory guide to resource-sensivity;
The community responsible for developing lexicons for Natural Language Processing (NLP) and Machine Readable Dictionaries (MRDs) started their ISO standardization activities in 2003. These activities resulted in the ISO standard - Lexical Markup Framework (LMF).After selecting and defining a common terminology, the LMF team had to identify the common notions shared by all lexicons in order to specify a common skeleton (called the core model) and understand the various requirements coming from different groups of users.The goals of LMF are to provide a common model for the creation and use of lexical resources, to manage the exchange of data between and among these resources, and to enable the merging of a large number of individual electronic resources to form extensive global electronic resources.The various types of individual instantiations of LMF can include monolingual, bilingual or multilingual lexical resources. The same specifications can be used for small and large lexicons, both simple and complex, as well as for both written and spoken lexical representations. The descriptions range from morphology, syntax and computational semantics to computer-assisted translation. The languages covered are not restricted to European languages, but apply to all natural languages.The LMF specification is now a success and numerous lexicon managers currently use LMF in different languages and contexts.This book starts with the historical context of LMF, before providing an overview of the LMF model and the Data Category Registry, which provides a flexible means for applying constants like /grammatical gender/ in a variety of different settings. It then presents concrete applications and experiments on real data, which are important for developers who want to learn about the use of LMF. Contents 1. LMF - Historical Context and Perspectives, Nicoletta Calzolari, Monica Monachini and Claudia Soria.2. Model Description, Gil Francopoulo and Monte George.3. LMF and the Data Category Registry: Principles and Application, Menzo Windhouwer and Sue Ellen Wright.4. Wordnet-LMF: A Standard Representation for Multilingual Wordnets, Piek Vossen, Claudia Soria and Monica Monachini.5. Prolmf: A Multilingual Dictionary of Proper Names and their Relations, Denis Maurel, Beatrice Bouchou-Markhoff.6. LMF for Arabic, Aida Khemakhem, Bilel Gargouri, Kais Haddar and Abdelmajid Ben Hamadou.7. LMF for a Selection of African Languages, Chantal Enguehard and Mathieu Mangeot.8. LMF and its Implementation in Some Asian Languages, Takenobu Tokunaga, Sophia Y.M. Lee, Virach Sornlertlamvanich, Kiyoaki Shirai, Shu-Kai Hsieh and Chu-Ren Huang.9. DUELME: Dutch Electronic Lexicon of Multiword Expressions, Jan Odijk.10. UBY-LMF - Exploring the Boundaries of Language-Independent Lexicon Models, Judith Eckle-Kohler, Iryna Gurevych, Silvana Hartmann, Michael Matuschek and Christian M. Meyer.11. Conversion of Lexicon-Grammar Tables to LMF: Application to French, eric Laporte, Elsa Tolone and Matthieu Constant.12. Collaborative Tools: From Wiktionary to LMF, for Synchronic and Diachronic Language Data, Thierry Declerck, Pirsoka Lendvai and Karlheinz Morth.13. LMF Experiments on Format Conversions for Resource Merging: Converters and Problems, Marta Villegas, Muntsa Padro and Nuria Bel.14. LMF as a Foundation for Servicized Lexical Resources, Yoshihiko Hayashi, Monica Monachini, Bora Savas, Claudia Soria and Nicoletta Calzolari.15. Creating a Serialization of LMF: The Experience of the RELISH Project, Menzo Windhouwer, Justin Petro, Irina Nevskaya, Sebastian Drude, Helen Aristar-Dry and Jost Gippert.16. Global Atlas: Proper Nouns, From Wikipedia to LMF, Gil Francopoulo, Frederic Marcoul, David Causse and Gregory Piparo.17. LMF in U.S. Government Language Resource Management, Monte George. About the Authors Gil Francopoulo works for Tagmatica (www.tagmatica.com), a company specializing in software development in the field of linguistics and documentation in the semantic web, in Paris, France, as well as for Spotter (www.spotter.com), a company specializing in media and social media analytics. |
You may like...
The Natural Language for Artificial…
Dioneia Motta Monte-Serrat, Carlo Cattani
Paperback
R2,767
Discovery Miles 27 670
Pragmatic Issues in Specialized…
Francesca Bianchi, Sara Gesuato
Paperback
R2,288
Discovery Miles 22 880
Foundation Models for Natural Language…
Gerhard PaaĆ, Sven Giesselbach
Hardcover
R884
Discovery Miles 8 840
Machine Learning for Biometrics…
Partha Pratim Sarangi, Madhumita Panda, …
Paperback
R2,570
Discovery Miles 25 700
Trends in E-Tools and Resources for…
Gloria Corpas Pastor, Isabel Duran Munoz
Hardcover
R3,527
Discovery Miles 35 270
Spelling and Writing Words - Theoretical…
Cyril Perret, Thierry Olive
Hardcover
R3,256
Discovery Miles 32 560
|