![]() |
![]() |
Your cart is empty |
||
Books > Language & Literature > Language & linguistics > Computational linguistics
The lexicon is now a major focus of research in computational linguistics and natural language processing (NLP), as more linguistic theories concentrate on the lexicon and as the acquisition of an adequate vocabulary has become the chief bottleneck in developing practical NLP systems. This collection describes techniques of lexical representation within a unification-based framework and their linguistic application, concentrating on the issue of structuring the lexicon using inheritance and defaults. Topics covered include typed feature structures, default unification, lexical rules, multiple inheritance and non-monotonic reasoning. The contributions describe both theoretical results and implemented languages and systems, including DATR, the Stuttgart TFS and ISSCO's ELU. This book arose out of a workshop on default inheritance in the lexicon organized as a part of the Esprit ACQUILEX project on computational lexicography. Besides the contributed papers mentioned above, it contains a detailed description of the ACQUILEX lexical knowledge base (LKB) system and its use in the representation of lexicons extracted semi-automatically from machine-readable dictionaries.
Semantic interpretation and the resolution of ambiguity presents an important advance in computer understanding of natural language. While parsing techniques have been greatly improved in recent years, the approach to semantics has generally improved in recent years, the approach to semantics has generally been ad hoc and had little theoretical basis. Graeme Hirst offers a new, theoretically motivated foundation for conceptual analysis by computer, and shows how this framework facilitates the resolution of lexical and syntactic ambiguities. His approach is interdisciplinary, drawing on research in computational linguistics, artificial intelligence, montague semantics, and cognitive psychology.
This book explores the relationship between online second language (L2) communicative activities and formal language learning. It provides empirical evidence of the scale of L2 English use online, investigating the forms most commonly used, the activities likely to cause discomfort and the challenges experienced by users, and takes a critical approach to the nature of language online beyond the paradigms of 'written' versus 'spoken'. The author explores the possibilities for language teaching practices that engage with and integrate learners' L2 English online use, not only to support it but to use it as input for classroom learning and to enhance and exploit its incidental learning outcomes. This book will be of interest to postgraduate students and researchers interested in computer-mediated communication, online discourse and Activity Theory, while language teachers will find the practical ideas for lesson content invaluable as they strive to create a successful language learning community.
Deep learning is revolutionizing how machine translation systems are built today. This book introduces the challenge of machine translation and evaluation - including historical, linguistic, and applied context -- then develops the core deep learning methods used for natural language applications. Code examples in Python give readers a hands-on blueprint for understanding and implementing their own machine translation systems. The book also provides extensive coverage of machine learning tricks, issues involved in handling various forms of data, model enhancements, and current challenges and methods for analysis and visualization. Summaries of the current research in the field make this a state-of-the-art textbook for undergraduate and graduate classes, as well as an essential reference for researchers and developers interested in other applications of neural methods in the broader field of human language processing.
This book is at the very heart of linguistics. It provides the theoretical and methodological framework needed to create a successful linguistic project. Potential applications of descriptive linguistics include spell-checkers, intelligent search engines, information extractors and annotators, automatic summary producers, automatic translators, and more. These applications have considerable economic potential, and it is therefore important for linguists to make use of these technologies and to be able to contribute to them. The author provides linguists with tools to help them formalize natural languages and aid in the building of software able to automatically process texts written in natural language (Natural Language Processing, or NLP). Computers are a vital tool for this, as characterizing a phenomenon using mathematical rules leads to its formalization. NooJ a linguistic development environment software developed by the author is described and practically applied to examples of NLP.
This case study-based textbook in multivariate analysis for advanced students in the humanities emphasizes descriptive, exploratory analyses of various types of datasets from a wide range of sub-disciplines, promoting the use of multivariate analysis and illustrating its wide applicability. Fields featured include, but are not limited to, historical agriculture, arts (music and painting), theology, and stylometrics (authorship issues). Most analyses are based on existing data, earlier analysed in published peer-reviewed papers. Four preliminary methodological and statistical chapters provide general technical background to the case studies. The multivariate statistical methods presented and illustrated include data inspection, several varieties of principal component analysis, correspondence analysis, multidimensional scaling, cluster analysis, regression analysis, discriminant analysis, and three-mode analysis. The bulk of the text is taken up by 14 case studies that lean heavily on graphical representations of statistical information such as biplots, using descriptive statistical techniques to support substantive conclusions. Each study features a description of the substantive background to the data, followed by discussion of appropriate multivariate techniques, and detailed results interpreted through graphical illustrations. Each study is concluded with a conceptual summary. Datasets in SPSS are included online.
Recent decades have seen a fundamental change and transformation in the commercialisation and popularisation of sports and sporting events. Corpus Approaches to the Language of Sports uses corpus resources to offer new perspectives on the language and discourse of this increasingly popular and culturally significant area of research. Bringing together a range of empirical studies from leading scholars, this book bridges the gap between quantitative corpus approaches and more qualitative, multimodal discourse methods. Covering a wide range of sports, including football, cycling and basketball, the linguistic aspects of sports language are analysed across different genres and contexts. Highlighting the importance of studying the language of sports alongside its accompanying audio-visual modes of communication, chapters draw on new digitised collections of language to fully describe and understand the complexities of communication through various channels. In doing so, Corpus Approaches to the Language of Sports not only offers exciting new insights into the language of sports but also extends the scope of corpus linguistics beyond traditional monomodal approaches to put multimodality firmly on the agenda.
Seit dem Entstehen der modernen Textlinguistik in den 1960er Jahren ist eine Vielzahl z.T. hoch spezialisierter Analyseansatze in diesem Bereich entwickelt worden, die auch in diversen Einfuhrungen schon aufbereitet worden sind. Anliegen dieses Arbeitsheftes ist es, Grundlagen linguistischer Textanalyse vorzustellen, wie sie insbesondere Studierende philologischer Facher bei der Analyse literarischer und anspruchsvoller Sachtexte benotigen. Textlinguistik wird dabei nicht als eine Sonderdisziplin der Sprachwissenschaft aufgefasst, die sich nur mit der "obersten" Beschreibungsebene befasst, sondern im Sinne der von Peter Hartmann konzipierten "verwendungsorientierten Sprachwissenschaft." Besonderer Wert wird darauf gelegt, die "neue" Textlinguistik auch in die Tradition fruherer Bemuhungen um den Gegenstand einzuordnen (Rhetorik, Hermeneutik, Literaturwissenschaft, vorstrukturalistische Grammatik). Das Schwergewicht der Darstellung liegt auf der mit vielen Beispielen angereicherten Erlauterung der vier zentralen Beschreibungsdimensionen: situativer Kontext, Funktion, Thema, sprachliche Gestalt. Hier werden nicht nur die den Textzusammenhalt gewahrleistenden Kohasionsmittel besprochen, sondern die Gesamtheit der sprachlichen Mittel, v.a. auf der Ebene von Lexik und Grammatik. Ziel ist es, die Verbindung zwischen Variationslinguistik und Textlinguistik zu verdeutlichen: Zu den Aufgaben der letzteren gehort es, die Soll- und Ist-Normen von Varietaten und Textsorten zu beschreiben."
In the not so distant future, we can expect a world where humans and robots coexist and interact with each other. For this to occur, we need to understand human traits, such as seeing, hearing, thinking, speaking, etc., and institute these traits in robots. The most essential feature necessary for robots to achieve is that of integrative multimedia understanding (IMU) which occurs naturally in humans. It allows us to assimilate pieces of information expressed through different modes such as speech, pictures, gestures, etc. The book describes how robots acquire traits like natural language understanding (NLU) as the central part of IMU. Mental image directed semantic theory (MIDST) is its core, and is based on the hypothesis that NLU is essentially the processing of mental image associated with natural language expressions, namely, mental-image based understanding (MBU). MIDST is intended to model omnisensory mental image in human and to afford a knowledge representation system in order for integrative management of knowledge subjective to cognitive mechanisms of intelligent entities such as humans and robots based on a mental image model visualized as 'Loci in Attribute Spaces' and its description language Lmd (mental image description language) to be employed for predicate logic with a systematic scheme for symbol-grounding. This language works as an interlingua among various kinds of information media, and has been applied to several versions of the intelligent system interlingual understanding model aiming at general system (IMAGES). Its latest version, i.e. conversation management system (CMS) simulates MBU and comprehends the user's intention through dialogue to find and solve problems, and finally, provides a response in text or animation. The book is aimed at researchers and students interested in artificial intelligence, robotics, and cognitive science. Based on philosophical considerations, the methodology will also have an appeal in linguistics, psychology, ontology, geography, and cartography. Key Features: Describes the methodology to provide robots with human-like capability of natural language understanding (NLU) as the central part of IMU Uses methodology that also relates to linguistics, psychology, ontology, geography, and cartography Examines current trends in machine translation
From tech giants to plucky startups, the world is full of companies boasting that they are on their way to replacing human interpreters, but are they right? Interpreters vs Machines offers a solid introduction to recent theory and research on human and machine interpreting, and then invites the reader to explore the future of interpreting. With a foreword by Dr Henry Liu, the 13th International Federation of Translators (FIT) President, and written by consultant interpreter and researcher Jonathan Downie, this book offers a unique combination of research and practical insight into the field of interpreting. Written in an innovative, accessible style with humorous touches and real-life case studies, this book is structured around the metaphor of playing and winning a computer game. It takes interpreters of all experience levels on a journey to better understand their own work, learn how computers attempt to interpret and explore possible futures for human interpreters. With five levels and split into 14 chapters, Interpreters vs Machines is key reading for all professional interpreters as well as students and researchers of Interpreting and Translation Studies, and those with an interest in machine interpreting.
Stress and accent are central, organizing features of grammar, but their precise nature continues to be a source of mystery and wonder. These issues come to the forefront in acquisition, where the tension between the abstract mental representations and the concrete physical manifestations of stress and accent is deeply reflected. Understanding the nature of the representations of stress and accent patterns, and understanding how stress and accent patterns are learned, informs all aspects of linguistic theory and language acquisition. These two themes - representation and acquisition - form the organizational backbone of this book. Each is addressed along different dimensions of stress and accent, including the position of an accent or stress within various prosodic domains and the acoustic dimensions along which the pronunciation of stress and accent may vary. The research presented in the book is multidisciplinary, encompassing theoretical linguistics, speech science, and computational and experimental research.
This book reviews ways to improve statistical machine speech translation between Polish and English. Research has been conducted mostly on dictionary-based, rule-based, and syntax-based, machine translation techniques. Most popular methodologies and tools are not well-suited for the Polish language and therefore require adaptation, and language resources are lacking in parallel and monolingual data. The main objective of this volume to develop an automatic and robust Polish-to-English translation system to meet specific translation requirements and to develop bilingual textual resources by mining comparable corpora.
Aiming at exemplifying the methodology of learner corpus profiling, this book describes salient features of Romanian Learner English. As a starting point, the volume offers a comprehensive presentation of the Romanian-English contrastive studies. Another innovative aspect of the book refers to the use of the first Romanian Corpus of Learner English, whose compilation is the object of a methodological discussion. In one of the main chapters, the book introduces the methodology of learner corpus profiling and compares it with existing approaches. The profiling approach is emphasised by corpus-based quantitative and qualitative investigations of Romanian Learner English. Part of the investigation is dedicated to the lexico-grammatical profiles of articles, prepositions and genitives. The frequency-based collocation analyses are integrated with error analyses and extended into error pattern samples. Furthermore, contrasting typical Romanian Learner English constructions with examples from the German and the Italian learner corpora opens the path to new contrastive interlanguage analyses.
Ruslan Mitkov's highly successful Oxford Handbook of Computational Linguistics has been substantially revised and expanded in this second edition. Alongside updated accounts of the topics covered in the first edition, it includes 17 new chapters on subjects such as semantic role-labelling, text-to-speech synthesis, translation technology, opinion mining and sentiment analysis, and the application of Natural Language Processing in educational and biomedical contexts, among many others. The volume is divided into four parts that examine, respectively: the linguistic fundamentals of computational linguistics; the methods and resources used, such as statistical modelling, machine learning, and corpus annotation; key language processing tasks including text segmentation, anaphora resolution, and speech recognition; and the major applications of Natural Language Processing, from machine translation to author profiling. The book will be an essential reference for researchers and students in computational linguistics and Natural Language Processing, as well as those working in related industries.
This textbook approaches second language acquisition from the perspective of generative linguistics. Roumyana Slabakova reviews and discusses paradigms and findings from the last thirty years of research in the field, focussing in particular on how the second or additional language is represented in the mind and how it is used in communication. The adoption and analysis of a specific model of acquisition, the Bottleneck Hypothesis, provides a unifying perspective. The book assumes some non-technical knowledge of linguistics, but important concepts are clearly introduced and defined throughout, making it a valuable resource not only for undergraduate and graduate students of linguistics, but also for researchers in cognitive science and language teachers.
The Lexicon provides an introduction to the study of words, their main properties, and how we use them to create meaning. It offers a detailed description of the organizing principles of the lexicon, and of the categories used to classify a wide range of lexical phenomena, including polysemy, meaning variation in composition, and the interplay with ontology, syntax, and pragmatics. Elisabetta Jezek uses empirical data from digitalized corpora and speakers' judgements, combined with the formalisms developed in the field of general and theoretical linguistics, to propose representations for each of these phenomena. The key feature of the book is that it merges theoretical accounts with lexicographic approaches and computational insights. Its clear structure and accessible approach make The Lexicon an ideal textbook for all students of linguistics-theoretical, applied, and computational-and a valuable resource for scholars and students of language in the fields of cognitive science and philosophy.
The content of this textbook is organized as a theory of language for the construction of talking robots. The main topic is the mechanism of natural language communication in both the speaker and the hearer. In the third edition the author has modernized the text, leaving the overview of traditional, theoretical, and computational linguistics, analytic philosophy of language, and mathematical complexity theory with their historical backgrounds intact. The format of the empirical analyses of English and German syntax and semantics has been adapted to current practice; and Chaps. 22-24 have been rewritten to focus more sharply on the construction of a talking robot.
This dictionary provides a full and authoritative guide to the
meanings of the terms, concepts, and theories employed in
pragmatics, the study of language in use.
In this book John A. Hawkins argues that major patterns of variation across languages are structured by general principles of efficiency in language use and communication. Evidence for these comes from languages permitting structural options from which selections are made in performance, e.g. between competing word orders and between relative clauses with a resumptive pronoun versus a gap. The preferences and patterns of performance within languages are reflected, he shows, in the fixed conventions and variation patterns across grammars, leading to a 'Performance-Grammar Correspondence Hypothesis'. Hawkins extends and updates the general theory that he laid out in Efficiency and Complexity in Grammars (OUP 2004): new areas of grammar and performance are discussed, new research findings are incorporated that test his earlier predictions, and new advances in the contributing fields of language processing, linguistic theory, historical linguistics, and typology are addressed. This efficiency approach to variation has far-reaching theoretical consequences relevant to many current issues in the language sciences. These include the notion of ease of processing and how to measure it, the role of processing in language change, the nature of language universals and their explanation, the theory of complexity, the relative strength of competing and cooperating principles, and the proper definition of fundamental grammatical notions such as 'dependency'. The book also offers a new typology of VO and OV languages and their correlating properties seen from this perspective, and a new typology of the noun phrase and of argument structure.
In this book John A. Hawkins argues that major patterns of variation across languages are structured by general principles of efficiency in language use and communication. Evidence for these comes from languages permitting structural options from which selections are made in performance, e.g. between competing word orders and between relative clauses with a resumptive pronoun versus a gap. The preferences and patterns of performance within languages are reflected, he shows, in the fixed conventions and variation patterns across grammars, leading to a 'Performance-Grammar Correspondence Hypothesis'. Hawkins extends and updates the general theory that he laid out in Efficiency and Complexity in Grammars (OUP 2004): new areas of grammar and performance are discussed, new research findings are incorporated that test his earlier predictions, and new advances in the contributing fields of language processing, linguistic theory, historical linguistics, and typology are addressed. This efficiency approach to variation has far-reaching theoretical consequences relevant to many current issues in the language sciences. These include the notion of ease of processing and how to measure it, the role of processing in language change, the nature of language universals and their explanation, the theory of complexity, the relative strength of competing and cooperating principles, and the proper definition of fundamental grammatical notions such as 'dependency'. The book also offers a new typology of VO and OV languages and their correlating properties seen from this perspective, and a new typology of the noun phrase and of argument structure.
This book adopts a corpus-based critical discourse analysis approach and examines a corpus of newspaper articles from Pakistani and Indian publications to gain comparative insights into the ideological construction of China's Belt and Road Initiative (BRI) and the China-Pakistan Economic Corridor (CPEC) within news discourses. This book contributes to the works on perceptions of BRI in English newspapers of India and Pakistan. A multi-billion-dollar project of BRI or the "One Belt One Road" (OBOR), CPEC symbolizes a vision for regional revival under China's economic leadership and clout. Propelled by the Chinese Premier's dream to revive the Chinese economy as well as to restructure and catalyze infrastructural development in Asia, BRI is aimed at connecting Asia via land and sea routes with Europe, Africa, and the Middle Eastern states.
This handbook explores multiple facets of the study of word classes, also known as parts of speech or lexical categories. These categories are of fundamental importance to linguistic theory and description, both formal and functional, and for both language-internal analyses and cross-linguistic comparison. The volume consists of five parts that investigate word classes from different angles. Chapters in the first part address a range of fundamental issues including diversity and unity in word classes around the world, categorization at different levels of structure, the distinction between lexical and functional words, and hybrid categories. Part II examines the treatment of word classes across a wide range of contemporary linguistic theories, such as Cognitive Grammar, Minimalist Syntax, and Lexical Functional Grammar, while the focus of Part III is on individual word classes, from major categories such as verb and noun to minor ones such as adpositions and ideophones. Part IV provides a number of cross-linguistic case studies, exploring word classes in families including Afroasiatic, Sinitic, Mayan, Austronesian, and in sign languages. Chapters in the final part of the book discuss word classes from the perspective of various sub-disciplines of linguistics, ranging from first and second language acquisition to computational and corpus linguistics. Together, the contributions showcase the importance of word classes for the whole discipline of linguistics, while also highlighting the many ongoing debates in the areas and outlining fruitful avenues for future research.
The two-volume set LNCS 13396 and 13397 constitutes revised selected papers from the CICLing 2018 conference which took place in Hanoi, Vietnam, in March 2018.The total of 67 papers presented in the two volumes was carefully reviewed and selected from 181 submissions. The focus of the conference was on following topics such as computational linguistics and intelligent text and speech processing and others. The papers are organized in the following topical sections: General, Author profiling and authorship attribution, social network analysis, Information retrieval, information extraction, Lexical resources, Machine translation, Morphology, syntax, Semantics and text similarity, Sentiment analysis, Syntax and parsing, Text categorization and clustering, Text generation, and Text mining.
This is an essential guide to using digital resources in the study of English language and linguistics. Assuming no prior experience, it introduces the fundamentals of online corpora and equips readers with the skills needed to search and interpret corpus data. Later chapters focus on specific elements of linguistic analysis, namely vocabulary, grammar, discourse and pronunciation. Examples from five major online corpora illustrate key issues to consider in corpus analysis, while case studies and activities help students get to grips with the wide range of resources that are available and select those that best suit their needs. Perfect for students of corpus linguistics and applied linguistics, this engaging and accessible guide opens the door to an ever-expanding world of online resources. It is also ideal for anyone who is curious about how the English language works and has a desire to explore its many written and spoken forms. New to this Edition: - Fully revised and updated throughout, incorporating the latest developments in corpus linguistics - Expanded material on corpora in teaching, contextualising corpus texts and critical discourse analysis
Interpreting Motion presents an integrated perspective on how language structures constrain concepts of motion and how the world shapes the way motion is linguistically expressed. Natural language allows for efficient communication of elaborate descriptions of movement without requiring a precise specification of the motion. Interpreting Motion is the first book to analyze the semantics of motion expressions in terms of the formalisms of qualitative spatial reasoning. It shows how motion descriptions in language are mapped to trajectories of moving entities based on qualitative spatio-temporal relationships. The authors provide an extensive discussion of prior research on spatial prepositions and motion verbs, devoting chapters to the compositional semantics of motion sentences, the formal representations needed for computers to reason qualitatively about time, space, and motion, and the methodology for annotating corpora with linguistic information in order to train computer programs to reproduce the annotation. The applications they illustrate include route navigation, the mapping of travel narratives, question-answering, image and video tagging, and graphical rendering of scenes from textual descriptions. The book is written accessibly for a broad scientific audience of linguists, cognitive scientists, computer scientists, and those working in fields such as artificial intelligence and geographic information systems. |
![]() ![]() You may like...
Foundation Models for Natural Language…
Gerhard PaaĂź, Sven Giesselbach
Hardcover
R935
Discovery Miles 9 350
Linguistic Inquiries into Donald…
Ulrike Schneider, Matthias Eitelmann
Hardcover
R4,241
Discovery Miles 42 410
Recent Developments in Fuzzy Logic and…
Shahnaz N. Shahbazova, Michio Sugeno, …
Hardcover
R6,324
Discovery Miles 63 240
|