![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Language & Literature > Language & linguistics > Computational linguistics
The main focus of this book is the investigation of linguistic variation in Spanish, considering spoken and written, specialised and non-specialised registers from a corpus linguistics approach and employing computational updated tools. The ten chapters represent a range of research on Spanish using a number of different corpora drawn from, amongst others, research articles, student writing, formal conversation and technical reports. A variety of methodologies are brought to bear upon these corpora including multi-dimensional and multi-register analysis, latent semantics and lexical bundles. This in-depth analysis of using Spanish corpora will be of interest to researchers in corpus linguistics or Spanish language. "Corpus and Discourse" series editors are: Wolfgang Teubert, University of Birmingham, and Michaela Mahlberg, Liverpool Hope University College. Editorial Board include: Frantisek Cermak (Prague), Susan Conrad (Portland), Geoffrey Leech (Lancaster), Elena Tognini-Bonelli (Lecce and TWC), Ruth Wodak (Lancaster and Vienna), Feng Zhiwei (Beijing). Corpus linguistics provides the methodology to extract meaning from texts. Taking as its starting point the fact that language is not a mirror of reality but lets us share what we know, believe and think about reality, it focuses on language as a social phenomenon, and makes visible the attitudes and beliefs expressed by the members of a discourse community. Consisting of both spoken and written language, discourse always has historical, social, functional, and regional dimensions. Discourse can be monolingual or multilingual, interconnected by translations. Discourse is where language and social studies meet. "The Corpus and Discourse" series consists of two strands. The first, "Research in Corpus and Discourse", features innovative contributions to various aspects of corpus linguistics and a wide range of applications, from language technology via the teaching of a second language to a history of mentalities. The second strand, "Studies in Corpus and Discourse", is comprised of key texts bridging the gap between social studies and linguistics. Although equally academically rigorous, this strand will be aimed at a wider audience of academics and postgraduate students working in both disciplines.
The key assumption in this text is that machine translation is not merely a mechanical process but in fact requires a high level of linguistic sophistication, as the nuances of syntax, semantics and intonation cannot always be conveyed by modern technology. The increasing dependence on artificial communication by private and corporate users makes this research area an invaluable element when teaching linguistic theory.
Describing the technologies to combine language resources flexibly as web services, this book provides valuable case studies for those who work in services computing, language resources, human-computer interaction (HCI), computer-supported cooperative work (CSCW), and service science. The authors have been operating the Language Grid, which wraps existing language resources as atomic language services and enables users to compose new services by combining them. From architecture level to service composition level, the book explains how to resolve infrastructural and operational difficulties in sharing and combining language resources, including interoperability of language service infrastructures, various types of language service policies, human services, and service failures.The research based on the authors' operating experiences of handling complicated issues such as intellectual property and interoperability of language resources contributes to exploitation of language resources as a service. On the other hand, both the analysis based on using services and the design of new services can bring significant results. A new style of multilingual communication supported by language services is worthy of analysis in HCI/CSCW, and the design process of language services is the focus of valuable case studies in service science. By using language resources in different ways based on the Language Grid, many activities are highly regarded by diverse communities. This book consists of four parts: (1) two types of language service platforms to interconnect language services across service grids, (2) various language service composition technologies that improve the reusability, efficiency, and accuracy of composite services, (3) research work and activities in creating language resources and services, and (4) various applications and tools for understanding and designing language services that well support intercultural collaboration.
Using a corpus of data drawn from naturally-occurring second
language conversations, this book explores the role of idiomaticity
in English as a native language, and its comparative role in
English as a lingua franca. Through examining how idiomaticity
enables first language learners to achieve a greater degree of
fluency, the book explores why idiomatic language poses such a
challenge for users of English as a lingua franca. The book puts
forward a new definition of competence and fluency within the
context of English as a lingua franca, concluding with an analysis
of practical implications for the lingua franca classroom.
This book showcases sixteen papers from the landmark 30th conference of the International Computer Archive of Modern and Medieval English (ICAME) held at Lancaster University in May 2009. The theme of the book 'looking back, moving forward' follows that of the conference where participants reflected on the extraordinary growth of corpus linguistics over three decades as well as looking ahead to yet further developments in the future. A separate volume, appearing as an e-publication in the VARIENG series from the University of Helsinki focuses on the methodological and historical dimensions of corpus linguistics. This volume features papers on present-day English and the recent history of English via the increasing availability of corpora covering the last hundred years or so of the language. Contributors to the volume study numerous topics and datasets including recent diachronic change, regional and new Englishes, learner corpora, Academic written English, parallel and translation corpora, corpora of popular music pop lyrics and computer-mediated communication. Overall the volume represents the state of the art in English corpus linguistics and a peek into the future directions for the field.
The book will appeal to scholars and advanced students of
morphology, syntax, computational linguistics and natural language
processing (NLP). It provides a critical and practical guide to
computational techniques for handling morphological and syntactic
phenomena, showing how these techniques have been used and modified
in practice.
English Corpora under Japanese Eyes is a fine collection of papers written in commemoration of the 10th anniversary of the Japan Association of English Corpus Studies (JAECS). Beginning with the overview of the field by Stig Johansson, an honorary member of the JAECS, the present volume shows the state-of-art in English corpus studies in Japan and demonstrates the creative uses of corpora in a wide range of research topics from studies drawing on large-scale general corpora, such as British National Corpus and the Bank of English, to studies based on more specific, historical, literary, learner and parallel corpora. The papers incorporated in this anthology are grouped into five sections: 1) Overview of corpus-based studies, 2) Corpus-based studies of contemporary English, 3) Historical and diachronic studies of English, 4) Corpus-based studies in English literature, 5) Corpus and English language teaching. This volume will inspire still further corpus exploitation in the broader field of the humanities.
Research monograph presenting a new approach to Computational Linguistics The ultimate goal of Computational Linguistics is to teach the computer to understand Natural Language. This research monograph presents a description of English according to algorithms which can be programmed into a computer to analyse natural language texts. The algorithmic approach uses series of instructions, written in Natural Language and organised in flow charts, with the aim of analysing certain aspects of the grammar of a sentence. One problem with text processing is the difficulty in distinguishing word forms that belong to parts of speech taken out of context. In order to solve this problem, Hristo Georgiev starts with the assumption that every word is either a verb or a non-verb. Form here he presents an algorithm which allows the computer to recognise parts of speech which to a human would be obvious though the meaning of the words. Emphasis for a computer is placed on verbs, nouns, participles and adjectives. English Algorithmic Grammar presents information for computers to recognise tenses, syntax, parsing, reference, and clauses. The final chapters of the book examine the further applications of an algorithmic approach to English grammar, and suggests ways in which the computer can be programmed to recognise meaning. This is an innovative, cutting-edge approach to computational linguistics that will be essential reading for academics researching computational linguistics, machine translation and natural language processing.
Corpus linguistics is often regarded as a methodology in its own right, but little attention has been given to the theoretical perspectives from which the subject can be approached. The present book contributes to filling this gap. Bringing together original contributions by internationally renowned authors, the chapters include coverage of the lexical priming theory, parole-linguistics, a four-part model of language system and language use, and the concept of local textual functions. The theoretical arguments are illustrated and complemented by case studies using data from large corpora such as the BNC, smaller purpose-built corpora, and Google searches. By presenting theoretical positions in corpus linguistics, "Text, Discourse, and Corpora" provides an essential overview for advanced undergraduate, postgraduate and academic readers. "Corpus and Discourse Series" editors are: Wolfgang Teubert, University of Birmingham, and Michaela Mahlberg, Liverpool Hope University College. Editorial Board: Frantisek Cermak (Prague), Susan Conrad (Portland), Geoffrey Leech (Lancaster), Elena Tognini-Bonelli (Lecce and TWC), Ruth Wodak (Lancaster and Vienna), and Feng Zhiwei (Beijing). Corpus linguistics provides the methodology to extract meaning from texts. Taking as its starting point the fact that language is not a mirror of reality but lets us share what we know, believe and think about reality, it focuses on language as a social phenomenon, and makes visible the attitudes and beliefs expressed by the members of a discourse community. Consisting of both spoken and written language, discourse always has historical, social, functional, and regional dimensions. Discourse can be monolingual or multilingual, interconnected by translations. Discourse is where language and social studies meet. "The Corpus and Discourse" series consists of two strands. The first, "Research in Corpus and Discourse", features innovative contributions to various aspects of corpus linguistics and a wide range of applications, from language technology via the teaching of a second language to a history of mentalities. The second strand, "Studies in Corpus and Discourse", is comprised of key texts bridging the gap between social studies and linguistics. Although equally academically rigorous, this strand will be aimed at a wider audience of academics and postgraduate students working in both disciplines.
This book presents multibiometric watermarking techniques for security of biometric data. This book also covers transform domain multibiometric watermarking techniques and their advantages and limitations. The authors have developed novel watermarking techniques with a combination of Compressive Sensing (CS) theory for the security of biometric data at the system database of the biometric system. The authors show how these techniques offer higher robustness, authenticity, better imperceptibility, increased payload capacity, and secure biometric watermarks. They show how to use the CS theory for the security of biometric watermarks before embedding into the host biometric data. The suggested methods may find potential applications in the security of biometric data at various banking applications, access control of laboratories, nuclear power stations, military base, and airports.
This book explains advanced theoretical and application-related issues in grammatical inference, a research area inside the inductive inference paradigm for machine learning. The first three chapters of the book deal with issues regarding theoretical learning frameworks; the next four chapters focus on the main classes of formal languages according to Chomsky's hierarchy, in particular regular and context-free languages; and the final chapter addresses the processing of biosequences. The topics chosen are of foundational interest with relatively mature and established results, algorithms and conclusions. The book will be of value to researchers and graduate students in areas such as theoretical computer science, machine learning, computational linguistics, bioinformatics, and cognitive psychology who are engaged with the study of learning, especially of the structure underlying the concept to be learned. Some knowledge of mathematics and theoretical computer science, including formal language theory, automata theory, formal grammars, and algorithmics, is a prerequisite for reading this book.
In the course of his career, Professor Halliday has continued to address the issue of the application of linguistic scholarship for computational and quantitative studies. The sixth volume in the collected works of Professor M.A.K. Halliday includes works that span the last five decades, covering such topics as machine translation: the early years; and probabilistic grammar. The last section of this volume includes discussion of recent collaborative efforts bringing together those working in systemic functional grammar, fuzzy logic and "intelligent computing," engaging in what Halliday refers to as computing with meaning. The Collected Works of M.A.K. Halliday is a series that brings together Halliday's publications in many branches of linguistics, both theoretical and applied (a distinction which he himself rejects), including grammar and semantics, discourse analysis and stylistics, phonology, sociolinguistics, computational linguistics, language education and child language development.
This book reports on an outstanding thesis that has significantly advanced the state-of-the-art in the automated analysis and classification of speech and music. It defines several standard acoustic parameter sets and describes their implementation in a novel, open-source, audio analysis framework called openSMILE, which has been accepted and intensively used worldwide. The book offers extensive descriptions of key methods for the automatic classification of speech and music signals in real-life conditions and reports on the evaluation of the framework developed and the acoustic parameter sets that were selected. It is not only intended as a manual for openSMILE users, but also and primarily as a guide and source of inspiration for students and scientists involved in the design of speech and music analysis methods that can robustly handle real-life conditions.
In this book, Stefan Th. Gries provides an overview on how quantitative corpus methods can provide insights to cognitive/usage-based linguistics and selected psycholinguistic questions. Topics include the corpus linguistics in general, its most important methodological tools, its statistical nature, and the relation of all these topics to past and current usage-based theorizing. Central notions discussed in detail include frequency, dispersion, context, and others in a variety of applications and case studies; four practice sessions offer short introductions of how to compute various corpus statistics with the open source programming language and environment R.
This book introduces Meaningful Purposive Interaction Analysis (MPIA) theory, which combines social network analysis (SNA) with latent semantic analysis (LSA) to help create and analyse a meaningful learning landscape from the digital traces left by a learning community in the co-construction of knowledge. The hybrid algorithm is implemented in the statistical programming language and environment R, introducing packages which capture - through matrix algebra - elements of learners' work with more knowledgeable others and resourceful content artefacts. The book provides comprehensive package-by-package application examples, and code samples that guide the reader through the MPIA model to show how the MPIA landscape can be constructed and the learner's journey mapped and analysed. This building block application will allow the reader to progress to using and building analytics to guide students and support decision-making in learning.
In this thesis, the author makes several contributions to the study of design of graphical materials. The thesis begins with a review of the relationship between design and aesthetics, and the use of mathematical models to capture this relationship. Then, a novel method for linking linguistic concepts to colors using the Latent Dirichlet Allocation Dual Topic Model is proposed. Next, the thesis studies the relationship between aesthetics and spatial layout by formalizing the notion of visual balance. Applying principles of salience and Gaussian mixture models over a body of about 120,000 aesthetically rated professional photographs, the author provides confirmation of Arnhem's theory about spatial layout. The thesis concludes with a description of tools to support automatically generating personalized design.
This book explores novel aspects of social robotics, spoken dialogue systems, human-robot interaction, spoken language understanding, multimodal communication, and system evaluation. It offers a variety of perspectives on and solutions to the most important questions about advanced techniques for social robots and chat systems. Chapters by leading researchers address key research and development topics in the field of spoken dialogue systems, focusing in particular on three special themes: dialogue state tracking, evaluation of human-robot dialogue in social robotics, and socio-cognitive language processing. The book offers a valuable resource for researchers and practitioners in both academia and industry whose work involves advanced interaction technology and who are seeking an up-to-date overview of the key topics. It also provides supplementary educational material for courses on state-of-the-art dialogue system technologies, social robotics, and related research fields.
This textbook examines empirical linguistics from a theoretical linguist's perspective. It provides both a theoretical discussion of what quantitative corpus linguistics entails and detailed, hands-on, step-by-step instructions to implement the techniques in the field. The statistical methodology and R-based coding from this book teach readers the basic and then more advanced skills to work with large data sets in their linguistics research and studies. Massive data sets are now more than ever the basis for work that ranges from usage-based linguistics to the far reaches of applied linguistics. This book presents much of the methodology in a corpus-based approach. However, the corpus-based methods in this book are also essential components of recent developments in sociolinguistics, historical linguistics, computational linguistics, and psycholinguistics. Material from the book will also be appealing to researchers in digital humanities and the many non-linguistic fields that use textual data analysis and text-based sensorimetrics. Chapters cover topics including corpus processing, frequencing data, and clustering methods. Case studies illustrate each chapter with accompanying data sets, R code, and exercises for use by readers. This book may be used in advanced undergraduate courses, graduate courses, and self-study.
All human speech has expression. It is part of the 'humanness' of speech, and is a quality listeners expect to find. Without expression, speech sounds lifeless and artificial. Remove expression, and what's left is the bare bones of the intended message, but none of the feelings which surround the message. The purpose of this book is to present research examining expressive content in speech with a view to simulating expression in computer speech. Human beings communicate expressively with each other in conversation: now in the computer age there is a perceived eed for machines to communicate expressively with humans in dialogue.
This contributed volume explores the achievements gained and the remaining puzzling questions by applying dynamical systems theory to the linguistic inquiry. In particular, the book is divided into three parts, each one addressing one of the following topics: 1) Facing complexity in the right way: mathematics and complexity 2) Complexity and theory of language 3) From empirical observation to formal models: investigation of specific linguistic phenomena, like enunciation, deixis, or the meaning of the metaphorical phrases The application of complexity theory to describe cognitive phenomena is a recent and very promising trend in cognitive science. At the time when dynamical approaches triggered a paradigm shift in cognitive science some decade ago, the major topic of research were the challenges imposed by classical computational approaches dealing with the explanation of cognitive phenomena like consciousness, decision making and language. The target audience primarily comprises researchers and experts in the field but the book may also be beneficial for graduate and post-graduate students who want to enter the field.
This book integrates current advances in biology, economics of information and linguistics research through applications using agent-based modeling and social network analysis to develop scenarios of communication and language emergence in the social aspects of biological communications. The book presents a model of communication emergence that can be applied both to human and non-human living organism networks. The model is based on economic concepts and individual behavior fundamental for the study of trust and reputation networks in social science, particularly in economics; it is also based on the theory of the emergence of norms and historical path dependence that has been influential in institutional economics. Also included are mathematical models and code for agent-based models to explore various scenarios of language evolution, as well as a computer application that explores language and communication in biological versus social organisms, and the emergence of various meanings and grammars in human networks. Emergence of Communication in Socio-Biological Networks offers both a completely novel approach to communication emergence and language evolution and provides a path for the reader to explore various scenarios of language and communication that are not constrained to the human networks alone. By illustrating how computational social science and the complex systems approach can incorporate multiple disciplines and offer an integrated theory-model approach to the evolution of language, the book will be of interest to researchers working with computational linguistics, mathematical linguistics, and complex systems.
This book offers an introduction to modern natural language processing using machine learning, focusing on how neural networks create a machine interpretable representation of the meaning of natural language. Language is crucially linked to ideas - as Webster's 1923 "English Composition and Literature" puts it: "A sentence is a group of words expressing a complete thought". Thus the representation of sentences and the words that make them up is vital in advancing artificial intelligence and other "smart" systems currently being developed. Providing an overview of the research in the area, from Bengio et al.'s seminal work on a "Neural Probabilistic Language Model" in 2003, to the latest techniques, this book enables readers to gain an understanding of how the techniques are related and what is best for their purposes. As well as a introduction to neural networks in general and recurrent neural networks in particular, this book details the methods used for representing words, senses of words, and larger structures such as sentences or documents. The book highlights practical implementations and discusses many aspects that are often overlooked or misunderstood. The book includes thorough instruction on challenging areas such as hierarchical softmax and negative sampling, to ensure the reader fully and easily understands the details of how the algorithms function. Combining practical aspects with a more traditional review of the literature, it is directly applicable to a broad readership. It is an invaluable introduction for early graduate students working in natural language processing; a trustworthy guide for industry developers wishing to make use of recent innovations; and a sturdy bridge for researchers already familiar with linguistics or machine learning wishing to understand the other.
In this book, leading researchers in morphology, syntax, language acquisition, psycholinguistics, and computational linguistics address central questions about the form and acquisition of analogy in grammar. What kinds of patterns do speakers select as the basis for analogical extension? What types of items are particularly susceptible or resistant to analogical pressures? At what levels do analogical processes operate and how do processes interact? What formal mechanisms are appropriate for modelling analogy? The novel synthesis of typological, theoretical, computational, and developmental paradigms in this volume brings us closer to answering these questions than ever before.
Cross-Disciplinary Advances in Applied Natural Language Processing: Issues and Approaches defines the role of ANLP within NLP, and alongside other disciplines such as linguistics, computer science, and cognitive science. The description also includes the categorization of current ANLP research, and examples of current research in ANLP. This book is a useful reference for teachers, students, and materials developers in fields spanning linguistics, computer science, and cognitive science.
This book presents the consolidated acoustic data for all phones in Standard Colloquial Bengali (SCB), commonly known as Bangla, a Bengali language used by 350 million people in India, Bangladesh, and the Bengali diaspora. The book analyzes the real speech of selected native speakers of the Bangla dialect to ensure that a proper acoustical database is available for the development of speech technologies. The acoustic data presented consists of averages and their normal spread, represented by the standard deviations of necessary acoustic parameters including e.g. formant information for multiple native speakers of both sexes. The study employs two important speech technologies:(1) text to speech synthesis (TTS) and (2) automatic speech recognition (ASR). The procedures, particularly those related to the use of technologies, are described in sufficient detail to enable researchers to use them to create technical acoustic databases for any other Indian dialect. The book offers a unique resource for scientists and industrial practitioners who are interested in the acoustic analysis and processing of Indian dialects to develop similar dialect databases of their own. |
You may like...
Linguistic Inquiries into Donald…
Ulrike Schneider, Matthias Eitelmann
Hardcover
R3,992
Discovery Miles 39 920
The Oxford Handbook of Information…
Caroline Fery, Shinichiro Ishihara
Hardcover
R4,569
Discovery Miles 45 690
Machine Learning for Biometrics…
Partha Pratim Sarangi, Madhumita Panda, …
Paperback
R2,570
Discovery Miles 25 700
Genres on the Web - Computational Models…
Alexander Mehler, Serge Sharoff, …
Hardcover
R4,212
Discovery Miles 42 120
Spelling and Writing Words - Theoretical…
Cyril Perret, Thierry Olive
Hardcover
R2,803
Discovery Miles 28 030
Artificial Intelligence for Healthcare…
Boris Galitsky, Saveli Goldberg
Paperback
R2,991
Discovery Miles 29 910
|