![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Language & Literature > Language & linguistics > Computational linguistics
The breadth and spread of corpus-assisted discourse studies (CADS) indicate its usefulness for exploring language use within a social context. However, its theoretical foundations, limitations, and its epistemological implications must be considered so that we can adjust our research designs accordingly. This Element focuses on important meta-level questions around epistemology, while also offering a compact guide to which corpus linguistic tools are available and how they can contribute to finding out more about discourse. This Element will appeal to researchers both new and experienced, both within the CADS community and beyond.
Dynamical Grammar explores the consequences for language acquisition, language evolution, and linguistic theory of taking the underlying architecture of the language faculty to be that of a complex adaptive dynamical system. It contains the first results of a new and complex model of language acquisition which the authors have developed to measure how far language input is reflected in language output and thereby get a better idea of just how far the human language faculty is hard-wired.
This book adopts a corpus-based critical discourse analysis approach and examines a corpus of newspaper articles from Pakistani and Indian publications to gain comparative insights into the ideological construction of China's Belt and Road Initiative (BRI) and the China-Pakistan Economic Corridor (CPEC) within news discourses. This book contributes to the works on perceptions of BRI in English newspapers of India and Pakistan. A multi-billion-dollar project of BRI or the "One Belt One Road" (OBOR), CPEC symbolizes a vision for regional revival under China's economic leadership and clout. Propelled by the Chinese Premier's dream to revive the Chinese economy as well as to restructure and catalyze infrastructural development in Asia, BRI is aimed at connecting Asia via land and sea routes with Europe, Africa, and the Middle Eastern states.
This case study-based textbook in multivariate analysis for advanced students in the humanities emphasizes descriptive, exploratory analyses of various types of datasets from a wide range of sub-disciplines, promoting the use of multivariate analysis and illustrating its wide applicability. Fields featured include, but are not limited to, historical agriculture, arts (music and painting), theology, and stylometrics (authorship issues). Most analyses are based on existing data, earlier analysed in published peer-reviewed papers. Four preliminary methodological and statistical chapters provide general technical background to the case studies. The multivariate statistical methods presented and illustrated include data inspection, several varieties of principal component analysis, correspondence analysis, multidimensional scaling, cluster analysis, regression analysis, discriminant analysis, and three-mode analysis. The bulk of the text is taken up by 14 case studies that lean heavily on graphical representations of statistical information such as biplots, using descriptive statistical techniques to support substantive conclusions. Each study features a description of the substantive background to the data, followed by discussion of appropriate multivariate techniques, and detailed results interpreted through graphical illustrations. Each study is concluded with a conceptual summary. Datasets in SPSS are included online.
The two-volume set LNCS 13396 and 13397 constitutes revised selected papers from the CICLing 2018 conference which took place in Hanoi, Vietnam, in March 2018.The total of 67 papers presented in the two volumes was carefully reviewed and selected from 181 submissions. The focus of the conference was on following topics such as computational linguistics and intelligent text and speech processing and others. The papers are organized in the following topical sections: General, Author profiling and authorship attribution, social network analysis, Information retrieval, information extraction, Lexical resources, Machine translation, Morphology, syntax, Semantics and text similarity, Sentiment analysis, Syntax and parsing, Text categorization and clustering, Text generation, and Text mining.
In the not so distant future, we can expect a world where humans and robots coexist and interact with each other. For this to occur, we need to understand human traits, such as seeing, hearing, thinking, speaking, etc., and institute these traits in robots. The most essential feature necessary for robots to achieve is that of integrative multimedia understanding (IMU) which occurs naturally in humans. It allows us to assimilate pieces of information expressed through different modes such as speech, pictures, gestures, etc. The book describes how robots acquire traits like natural language understanding (NLU) as the central part of IMU. Mental image directed semantic theory (MIDST) is its core, and is based on the hypothesis that NLU is essentially the processing of mental image associated with natural language expressions, namely, mental-image based understanding (MBU). MIDST is intended to model omnisensory mental image in human and to afford a knowledge representation system in order for integrative management of knowledge subjective to cognitive mechanisms of intelligent entities such as humans and robots based on a mental image model visualized as 'Loci in Attribute Spaces' and its description language Lmd (mental image description language) to be employed for predicate logic with a systematic scheme for symbol-grounding. This language works as an interlingua among various kinds of information media, and has been applied to several versions of the intelligent system interlingual understanding model aiming at general system (IMAGES). Its latest version, i.e. conversation management system (CMS) simulates MBU and comprehends the user's intention through dialogue to find and solve problems, and finally, provides a response in text or animation. The book is aimed at researchers and students interested in artificial intelligence, robotics, and cognitive science. Based on philosophical considerations, the methodology will also have an appeal in linguistics, psychology, ontology, geography, and cartography. Key Features: Describes the methodology to provide robots with human-like capability of natural language understanding (NLU) as the central part of IMU Uses methodology that also relates to linguistics, psychology, ontology, geography, and cartography Examines current trends in machine translation
The two-volume set LNCS 13396 and 13397 constitutes revised selected papers from the CICLing 2018 conference which took place in Hanoi, Vietnam, in March 2018.The total of 67 papers presented in the two volumes was carefully reviewed and selected from 181 submissions. The focus of the conference was on following topics such as computational linguistics and intelligent text and speech processing and others. The papers are organized in the following topical sections: General, Author profiling and authorship attribution, social network analysis, Information retrieval, information extraction, Lexical resources, Machine translation, Morphology, syntax, Semantics and text similarity, Sentiment analysis, Syntax and parsing, Text categorization and clustering, Text generation, and Text mining.
This Element explores approaches to locating and examining social identity in corpora with and without the aid of demographic metadata. This is a key concern in corpus-aided studies of language and identity, and this Element sets out to explore the main challenges and affordances associated with either approach and to discern what either approach can (and cannot) show. It describes two case studies which each compare two approaches to social identity variables - sex and age - in a corpus of 14-million words of patient comments about NHS cancer services in England. The first approach utilises demographic tags to group comments according to patients' sex/age while the second involves categorising cases where patients disclose their sex/age in their comments. This Element compares the findings from either approach, with the approaches themselves being critically discussed in terms of their implications for corpus-aided studies of language and identity.
This book is an introduction to the linguistic concepts of argumentation relevant for argument mining, an important research and development activity which can be viewed as a highly complex form of information retrieval, requiring high-level natural language processing technology. While the first four chapters develop the linguistic and conceptual aspects of argument expression, the last four are devoted to their application to argument mining. These chapters investigate the facets of argument annotation, as well as argument mining system architectures and evaluation. How annotations may be used to develop linguistic data and how to train learning algorithms is outlined. A simple implementation is then proposed. The book ends with an analysis of non-verbal argumentative discourse. Argument Mining is an introductory book for engineers or students of linguistics, artificial intelligence and natural language processing. Most, if not all, the concepts of argumentation crucial for argument mining are carefully introduced and illustrated in a simple manner.
This open access book introduces Vector semantics, which links the formal theory of word vectors to the cognitive theory of linguistics. The computational linguists and deep learning researchers who developed word vectors have relied primarily on the ever-increasing availability of large corpora and of computers with highly parallel GPU and TPU compute engines, and their focus is with endowing computers with natural language capabilities for practical applications such as machine translation or question answering. Cognitive linguists investigate natural language from the perspective of human cognition, the relation between language and thought, and questions about conceptual universals, relying primarily on in-depth investigation of language in use. In spite of the fact that these two schools both have 'linguistics' in their name, so far there has been very limited communication between them, as their historical origins, data collection methods, and conceptual apparatuses are quite different. Vector semantics bridges the gap by presenting a formal theory, cast in terms of linear polytopes, that generalizes both word vectors and conceptual structures, by treating each dictionary definition as an equation, and the entire lexicon as a set of equations mutually constraining all meanings.
The book covers theoretical work, approaches, applications, and techniques for computational models of information, language, and reasoning. Computational and technological developments that incorporate natural language are proliferating. Adequate coverage of natural language processing in artificial intelligence encounters problems on developments of specialized computational approaches and algorithms. Many difficulties are due to ambiguities in natural language and dependency of interpretations on contexts and agents. Classical approaches proceed with relevant updates, and new developments emerge in theories of formal and natural languages, computational models of information and reasoning, and related computerized applications. Its focus is on computational processing of human language and relevant medium languages, which can be theoretically formal, or for programming and specification of computational systems. The goal is to promote intelligent natural language processing, along with models of computation, language, reasoning, and other cognitive processes.
This open access book provides an in-depth description of the EU project European Language Grid (ELG). Its motivation lies in the fact that Europe is a multilingual society with 24 official European Union Member State languages and dozens of additional languages including regional and minority languages. The only meaningful way to enable multilingualism and to benefit from this rich linguistic heritage is through Language Technologies (LT) including Natural Language Processing (NLP), Natural Language Understanding (NLU), Speech Technologies and language-centric Artificial Intelligence (AI) applications. The European Language Grid provides a single umbrella platform for the European LT community, including research and industry, effectively functioning as a virtual home, marketplace, showroom, and deployment centre for all services, tools, resources, products and organisations active in the field. Today the ELG cloud platform already offers access to more than 13,000 language processing tools and language resources. It enables all stakeholders to deposit, upload and deploy their technologies and datasets. The platform also supports the long-term objective of establishing digital language equality in Europe by 2030 - to create a situation in which all European languages enjoy equal technological support. This is the very first book dedicated to Language Technology and NLP platforms. Cloud technology has only recently matured enough to make the development of a platform like ELG feasible on a larger scale. The book comprehensively describes the results of the ELG project. Following an introduction, the content is divided into four main parts: (I) ELG Cloud Platform; (II) ELG Inventory of Technologies and Resources; (III) ELG Community and Initiative; and (IV) ELG Open Calls and Pilot Projects.
Corpus linguistics continues to be a vibrant methodology applied across highly diverse fields of research in the language sciences. With the current steep rise in corpus sizes, computational power, statistical literacy and multi-purpose software tools, and inspired by neighbouring disciplines, approaches have diversified to an extent that calls for an intensification of the accompanying critical debate. Bringing together a team of leading experts, this book follows a unique design, comparing advanced methods and approaches current in corpus linguistics, to stimulate reflective evaluation and discussion. Each chapter explores the strengths and weaknesses of different datasets and techniques, presenting a case study and allowing readers to gauge methodological options in practice. Contributions also provide suggestions for further reading, and data and analysis scripts are included in an online appendix. This is an important and timely volume, and will be essential reading for any linguist interested in corpus-linguistic approaches to variation and change.
The practices of visual artists can never be decontextualised from language. Firstly, artists are constantly in dialogue with their peers, dealers, critics, and audiences about their creative activities and these interactions impact on the work they produce. Secondly, artists' conceptualisations of what artistic practice encompasses are always shaped by wider social discourses. These discourses, however, and their manifestation in the language of everyday life are subject to continual change, and potentially reshape the way that artists conceptualise their practices. Using a 235,000-word diachronic corpus developed from artists' interviews and statements, this Element investigates shifts in artists' use of language to conceptualise their art practice from 1950 to 2019. It then compares these shifts to see if they align with changes in the wider English lexicon and whether there might be a relationship between everyday language change and the aesthetic and conceptual developments that take place in the art world.
The book features recent attempts to construct corpora for specific purposes - e.g. multifactorial Dutch (parallel), Geasy Easy Language Corpus (intralingual), HK LegCo interpreting corpus - and showcases sophisticated and innovative corpus analysis methods. It proposes new approaches to address classical themes - i.e. translation pedagogy, translation norms and equivalence, principles of translation - and brings interdisciplinary perspectives - e.g. contrastive linguistics, cognition and metaphor studies - to cast new light. It is a timely reference for the researchers as well as postgraduate students who are interested in the applications of corpus technology to solving translation and interpreting problems.
Corpus Linguistics has revolutionised the world of language study and is an essential component of work in Applied Linguistics. This book, now in its second edition, provides a thorough introduction to all the key research issues in Corpus Linguistics, from the point of view of Applied Linguistics. The field has progressed a great deal since the first edition, so this edition has been completely rewritten to reflect these advances, whilst still maintaining the emphasis on hands-on corpus research of the first edition. It includes chapters on qualitative and quantitative research, applications in language teaching, discourse studies, and beyond. It also includes an extensive discussion of the place of Corpus Linguistics in linguistic theory, and provides numerous detailed examples of corpus studies throughout. Providing an accessible but thorough grounding to the fascinating, fast-moving field of Corpus Linguistics, this book is essential reading for the student and the researcher alike.
Corpora are ubiquitous in linguistic research, yet to date, there has been no consensus on how to conceptualize corpus representativeness and collect corpus samples. This pioneering book bridges this gap by introducing a conceptual and methodological framework for corpus design and representativeness. Written by experts in the field, it shows how corpora can be designed and built in a way that is both optimally suited to specific research agendas, and adequately representative of the types of language use in question. It considers questions such as 'what types of texts should be included in the corpus?', and 'how many texts are required?' - highlighting that the degree of representativeness rests on the dual pillars of domain considerations and distribution considerations. The authors introduce, explain, and illustrate all aspects of this corpus representativeness framework in a step-by-step fashion, using examples and activities to help readers develop practical skills in corpus design and evaluation.
Corpus analysis can be expanded and scaled up by incorporating computational methods from natural language processing. This Element shows how text classification and text similarity models can extend our ability to undertake corpus linguistics across very large corpora. These computational methods are becoming increasingly important as corpora grow too large for more traditional types of linguistic analysis. We draw on five case studies to show how and why to use computational methods, ranging from usage-based grammar to authorship analysis to using social media for corpus-based sociolinguistics. Each section is accompanied by an interactive code notebook that shows how to implement the analysis in Python. A stand-alone Python package is also available to help readers use these methods with their own data. Because large-scale analysis introduces new ethical problems, this Element pairs each new methodology with a discussion of potential ethical implications.
This work presents a discourse-aware Text Simplification approach that splits and rephrases complex English sentences within the semantic context in which they occur. Based on a linguistically grounded transformation stage, complex sentences are transformed into shorter utterances with a simple canonical structure that can be easily analyzed by downstream applications. To avoid breaking down the input into a disjointed sequence of statements that is difficult to interpret, the author incorporates the semantic context between the split propositions in the form of hierarchical structures and semantic relationships, thus generating a novel representation of complex assertions that puts a semantic layer on top of the simplified sentences. In a second step, she leverages the semantic hierarchy of minimal propositions to improve the performance of Open IE frameworks. She shows that such systems benefit in two dimensions. First, the canonical structure of the simplified sentences facilitates the extraction of relational tuples, leading to an improved precision and recall of the extracted relations. Second, the semantic hierarchy can be leveraged to enrich the output of existing Open IE approaches with additional meta-information, resulting in a novel lightweight semantic representation for complex text data in the form of normalized and context-preserving relational tuples.
This book brings together selected revised papers representing a multidisciplinary approach to language, music, and gesture, as well as their interaction. Among the number of multidisciplinary and comparative studies of the structure and organization of language and music, the presented book broadens the scope with the inclusion of gesture problems in the analyzed spectrum. A unique feature of the presented collection is that the papers, compiled in one volume, allow readers to see similarities and differences in gesture as an element of non-verbal communication and gesture as the main element of dance. In addition to enhancing the analysis, the data on the perception and comprehension of speech, music, and dance in regard to both their functioning in a natural situation and their reflection in various forms of performing arts makes this collection extremely useful for those who are interested in human cognitive abilities and performing skills. The book begins with a philosophical overview of recent neurophysiological studies reflecting the complexity of higher cognitive functions, which references the idea of the baroque style in art being neither linear nor stable. The following papers are allocated into 5 sections. The papers of the section "Language-Music-Gesture As Semiotic Systems" discuss the issues of symbolic and semiotic aspects of language, music, and gesture, including from the perspective of their notation. This is followed by the issues of "Language-Music-Gesture Onstage" and interaction within the idea of the "World as a Text." The papers of "Teaching Language and Music" present new teaching methods that take into account the interaction of all the cognitive systems examined. The papers of the last two sections focus on issues related primarily to language: The section "Verbalization Of Music And Gesture" considers the problem of describing musical text and non-verbal behavior with language, and papers in the final section "Emotions In Linguistics And Ai-Communication Systems" analyze the ways of expressing emotions in speech and the problems of organizing emotional communication with computer agents.
When viewed through a political lens, the act of defining terms in natural language arguably transforms knowledge into values. This unique volume explores how corporate, military, academic, and professional values shaped efforts to define computer terminology and establish an information engineering profession as a precursor to what would become computer science. As the Cold War heated up, U.S. federal agencies increasingly funded university researchers and labs to develop technologies, like the computer, that would ensure that the U.S. maintained economic prosperity and military dominance over the Soviet Union. At the same time, private corporations saw opportunities for partnering with university labs and military agencies to generate profits as they strengthened their business positions in civilian sectors. They needed a common vocabulary and principles of streamlined communication to underpin the technology development that would ensure national prosperity and military dominance. investigates how language standardization contributed to the professionalization of computer science as separate from mathematics, electrical engineering, and physics examines traditions of language standardization in earlier eras of rapid technology development around electricity and radio highlights the importance of the analogy of "the computer is like a human" to early explanations of computer design and logic traces design and development of electronic computers within political and economic contexts foregrounds the importance of human relationships in decisions about computer design This in-depth humanistic study argues for the importance of natural language in shaping what people come to think of as possible and impossible relationships between computers and humans. The work is a key reference in the history of technology and serves as a source textbook on the human-level history of computing. In addition, it addresses those with interests in sociolinguistic questions around technology studies, as well as technology development at the nexus of politics, business, and human relations.
This book addresses the research, analysis, and description of the methods and processes that are used in the annotation and processing of language corpora in advanced, semi-advanced, and non-advanced languages. It provides the background information and empirical data needed to understand the nature and depth of problems related to corpus annotation and text processing and shows readers how the linguistic elements found in texts are analyzed and applied to develop language technology systems and devices. As such, it offers valuable insights for researchers, educators, and students of linguistics and language technology.
This book presents a method of linking the ordered structure of the cosmos with human thoughts: the theory of language holography. In the view presented here, the cosmos is in harmony with the human body and language, and human thoughts are holographic with the cosmos at the level of language. In a word, the holographic relation is nothing more than the bridge by means of which Guanlian Qian connects the cosmos, human, and language. This is a vitally important contribution to linguistic and philosophical studies that cannot be ignored. The book has two main focus areas: outer language holography and inner language holography. These two areas constitute the core of the dynamic and holistic view put forward in the theory of language holography. The book's main properties can be summarized into the following points: First and foremost, it is a book created in toto by a Chinese scholar devoted to pragmatics, theoretical linguistics, and philosophy of language. Secondly, the book was accepted by a top Chinese publisher and was republished the second year, which reflected its value and appeal. Thirdly, in terms of writing style, the book is characterized by succinctness and logic. As a result, it reads fluidly and smoothly without redundancies, which is not that common in linguistic or even philosophical works. Lastly, as stated by the author in the introduction, "Creation is the development of previous capacities, but it is also the generation of new ones"; this book can be said to put this concept into practice. Overall, the book offers a unique resource to readers around the world who want to know more about the truly original and innovative studies of language in Chinese academia.
This book presents a theoretical study on aspect in Chinese, including both situation and viewpoint aspects. Unlike previous studies, which have largely classified linguistic units into different situation types, this study defines a set of ontological event types that are conceptually universal and on the basis of which different languages employ various linguistic devices to describe such events. To do so, it focuses on a particular component of events, namely the viewpoint aspect. It includes and discusses a wealth of examples to show how such ontological events are realized in Chinese. In addition, the study discusses how Chinese modal verbs and adverbs affect the distribution of viewpoint aspects associated with certain situation types. In turn, the book demonstrates how the proposed linguistic theory can be used in a computational context. Simply identifying events in terms of the verbs and their arguments is insufficient for real situations such as understanding the factivity and the logical/temporal relations between events. The proposed framework offers the possibility of analyzing events in Chinese text, yielding deep semantic information.
The two-volume proceedings, LNCS 13249 and 13250, constitutes the thoroughly refereed post-workshop proceedings of the 22nd Chinese Lexical Semantics Workshop, CLSW 2021, held in Nanjing, China in May 2021. The 68 full papers and 4 short papers were carefully reviewed and selected from 261 submissions. They are organized in the following topical sections: Lexical Semantics and General Linguistics; Natural Language Processing and Language Computing; Cognitive Science and Experimental Studies; Lexical Resources and Corpus Linguistics. |
You may like...
Machine Learning for Biometrics…
Partha Pratim Sarangi, Madhumita Panda, …
Paperback
R2,570
Discovery Miles 25 700
Artificial Intelligence for Healthcare…
Boris Galitsky, Saveli Goldberg
Paperback
R2,991
Discovery Miles 29 910
The Natural Language for Artificial…
Dioneia Motta Monte-Serrat, Carlo Cattani
Paperback
R2,767
Discovery Miles 27 670
Trends in E-Tools and Resources for…
Gloria Corpas Pastor, Isabel Duran Munoz
Hardcover
R3,025
Discovery Miles 30 250
Foundation Models for Natural Language…
Gerhard PaaĆ, Sven Giesselbach
Hardcover
R884
Discovery Miles 8 840
The Temporal Structure of Multimodal…
Laszlo Hunyadi, Istvan Szekrenyes
Hardcover
R2,653
Discovery Miles 26 530
|