![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Language & Literature > Language & linguistics > Computational linguistics
The description, automatic identification and further processing of web genres is a novel field of research in computational linguistics, NLP and related areas such as text-technology, digital humanities and web mining. One of the driving forces behind this research is the idea of genre-enabled search engines which enable users to additionally specify web genres that the documents to be retrieved should comply with (e.g., personal homepage, weblog, scientific article etc.). This book offers a thorough foundation of this upcoming field of research on web genres and document types in web-based social networking. It provides theoretical foundations of web genres, presents corpus linguistic approaches to their analysis and computational models for their classification. This includes research in the areas of web genre identification, web genre modelling and related fields such as genres and registers in web based communication social software-based document networks web genre ontologies and classification schemes text-technological models of web genres web content, structure and usage mining web genre classification web as corpus. The book addresses researchers who want to become acquainted with theoretical developments, computational models and their empirical evaluation in this field of research. It also addresses researchers who are interested in standards for the creation of corpora of web documents. Thus, the book concerns readers from many disciplines such as corpus linguistics, computational linguistics, text-technology and computer science.
th This volume is dedicated to Dov Gabbay who celebrated his 50 birthday in October 1995. Dov is one of the most outstanding and most productive researchers we have ever met. He has exerted a profound influence in major fields of logic, linguistics and computer science. His contributions in the areas of logic, language and reasoning are so numerous that a comprehensive survey would already fill half of this book. Instead of summarizing his work we decided to let him speak for himself. Sitting in a car on the way to Amsterdam airport he gave an interview to Jelle Gerbrandy and Anne-Marie Mineur. This recorded conversation with him, which is included gives a deep insight into his motivations and into his view of the world, the Almighty and, of course, the role of logic. In addition, this volume contains a partially annotated bibliography of his main papers and books. The length of the bibliography and the broadness of the topics covered there speaks for itself.
Reversible grammar allows computational models to be built that are equally well suited for the analysis and generation of natural language utterances. This task can be viewed from very different perspectives by theoretical and computational linguists, and computer scientists. The papers in this volume present a broad range of approaches to reversible, bi-directional, and non-directional grammar systems that have emerged in recent years. This is also the first collection entirely devoted to the problems of reversibility in natural language processing. Most papers collected in this volume are derived from presentations at a workshop held at the University of California at Berkeley in the summer of 1991 organised under the auspices of the Association for Computational Linguistics. This book will be a valuable reference to researchers in linguistics and computer science with interests in computational linguistics, natural language processing, and machine translation, as well as in practical aspects of computability.
Trajectories through Knowledge Space: A Dynamic Framework for Machine Comprehension provides an overview of many of the main ideas of connectionism (neural networks) and probabilistic natural language processing. Several areas of common overlap between these fields are described in which each community can benefit from the ideas and techniques of the other. The author's perspective on comprehension pulls together the most significant research of the last ten years and illustrates how we can move more forward onto the next level of intelligent text processing systems. A central focus of the book is the development of a framework for comprehension connecting research themes from cognitive psychology, cognitive science, corpus linguistics and artificial intelligence. The book proposes a new architecture for semantic memory, providing a framework for addressing the problem of how to represent background knowledge in a machine. This architectural framework supports a computational model of comprehension.Trajectories through Knowledge Space: A Dynamic Framework for Machine Comprehension is an excellent reference for researchers and professionals, and may be used as an advanced text for courses on the topic.
1. Structuralist Versus Analogical Descriptions ONE important purpose of this book is to compare two completely dif ferent approaches to describing language. The first of these approaches, commonly called stnlctllralist, is the traditional method for describing behavior. Its methods are found in many diverse fields - from biological taxonomy to literary criticism. A structuralist description can be broadly characterized as a system of classification. The fundamental question that a structuralist description attempts to answer is how a general contextual space should be partitioned. For each context in the partition, a rule is defined. The rule either specifies the behavior of that context or (as in a taxonomy) assigns a name to that context. Structuralists have implicitly assumed that descriptions of behavior should not only be correct, but should also minimize the number of rules and permit only the simplest possible contextual specifications. It turns out that these intuitive notions can actually be derived from more fundamental statements about the uncertainty of rule systems. Traditionally, linguistic analyses have been based on the idea that a language is a system of rules. Saussure, of course, is well known as an early proponent of linguistic structuralism, as exemplified by his characterization of language as "a self-contained whole and principle of classification" (Saussure 1966:9). Yet linguistic structuralism did not originate with Saussure - nor did it end with "American structuralism.""
Speech--to--Speech Translation: a Massively Parallel Memory-Based Approach describes one of the world's first successful speech--to--speech machine translation systems. This system accepts speaker-independent continuous speech, and produces translations as audio output. Subsequent versions of this machine translation system have been implemented on several massively parallel computers, and these systems have attained translation performance in the milliseconds range. The success of this project triggered several massively parallel projects, as well as other massively parallel artificial intelligence projects throughout the world. Dr. Hiroaki Kitano received the distinguished 'Computers and Thought Award' from the International Joint Conferences on Artificial Intelligence in 1993 for his work in this area, and that work is reported in this book.
Aiming at exemplifying the methodology of learner corpus profiling, this book describes salient features of Romanian Learner English. As a starting point, the volume offers a comprehensive presentation of the Romanian-English contrastive studies. Another innovative aspect of the book refers to the use of the first Romanian Corpus of Learner English, whose compilation is the object of a methodological discussion. In one of the main chapters, the book introduces the methodology of learner corpus profiling and compares it with existing approaches. The profiling approach is emphasised by corpus-based quantitative and qualitative investigations of Romanian Learner English. Part of the investigation is dedicated to the lexico-grammatical profiles of articles, prepositions and genitives. The frequency-based collocation analyses are integrated with error analyses and extended into error pattern samples. Furthermore, contrasting typical Romanian Learner English constructions with examples from the German and the Italian learner corpora opens the path to new contrastive interlanguage analyses.
Connection science is a new information-processing paradigm which attempts to imitate the architecture and process of the brain, and brings together researchers from disciplines as diverse as computer science, physics, psychology, philosophy, linguistics, biology, engineering, neuroscience and AI. Work in Connectionist Natural Language Processing (CNLP) is now expanding rapidly, yet much of the work is still only available in journals, some of them quite obscure. To make this research more accessible this book brings together an important and comprehensive set of articles from the journal CONNECTION SCIENCE which represent the state of the art in Connectionist natural language processing; from speech recognition to discourse comprehension. While it is quintessentially Connectionist, it also deals with hybrid systems, and will be of interest to both theoreticians as well as computer modellers. Range of topics covered: Connectionism and Cognitive Linguistics Motion, Chomsky's Government-binding Theory Syntactic Transformations on Distributed Representations Syntactic Neural Networks A Hybrid Symbolic/Connectionist Model for Understanding of Nouns Connectionism and Determinism in a Syntactic Parser Context Free Grammar Recognition Script Recognition with Hierarchical Feature Maps Attention Mechanisms in Language Script-Based Story Processing A Connectionist Account of Similarity in Vowel Harmony Learning Distributed Representations Connectionist Language Users Representation and Recognition of Temporal Patterns A Hybrid Model of Script Generation Networks that Learn about Phonological Features Pronunciation in Text-to-Speech Systems
Methods for studying writing processes have significantly developed over the last two decades. The rapid development of software tools which support the collection together with the display and analysis of writing process data and new input from various neighboring disciplines contribute to an increasingly detailed knowledge acquisition about the complex cognitive processes of writing. This volume, which focuses on research methods, mixed methods designs, conceptual considerations of writing process research, interdisciplinary research influences and the application of research methods in educational settings, provides an insight into the current status of the methodological development of writing process research in Europe.
Parsing technology traditionally consists of two branches, which correspond to the two main application areas of context-free grammars and their generalizations. Efficient deterministic parsing algorithms have been developed for parsing programming languages, and quite different algorithms are employed for analyzing natural language. The Functional Treatment of Parsing provides a functional framework within which the different traditional techniques are restated and unified. The resulting theory provides new recursive implementations of parsers for context-free grammars. The new implementations, called recursive ascent parsers, avoid explicit manipulation of parse stacks and parse matrices, and are in many ways superior to conventional implementations. They are applicable to grammars for programming languages as well as natural languages. The book has been written primarily for students and practitioners of parsing technology. With its emphasis on modern functional methods, however, the book will also be of benefit to scientists interested in functional programming. The Functional Treatment of Parsing is an excellent reference and can be used as a text for a course on the subject.
The Generalized LR parsing algorithm (some call it "Tomita's algorithm") was originally developed in 1985 as a part of my Ph.D thesis at Carnegie Mellon University. When I was a graduate student at CMU, I tried to build a couple of natural language systems based on existing parsing methods. Their parsing speed, however, always bothered me. I sometimes wondered whether it was ever possible to build a natural language parser that could parse reasonably long sentences in a reasonable time without help from large mainframe machines. At the same time, I was always amazed by the speed of programming language compilers, because they can parse very long sentences (i.e., programs) very quickly even on workstations. There are two reasons. First, programming languages are considerably simpler than natural languages. And secondly, they have very efficient parsing methods, most notably LR. The LR parsing algorithm first precompiles a grammar into an LR parsing table, and at the actual parsing time, it performs shift-reduce parsing guided deterministically by the parsing table. So, the key to the LR efficiency is the grammar precompilation; something that had never been tried for natural languages in 1985. Of course, there was a good reason why LR had never been applied for natural languages; it was simply impossible. If your context-free grammar is sufficiently more complex than programming languages, its LR parsing table will have multiple actions, and deterministic parsing will be no longer possible.
The volume brings together a selection of invited articles and papers presented at the 4th International CILC Conference held in Jaen, Spain, in March 2012. The chapters describe English using a range of corpora and other resources. There are two parts, one dealing with diachronic research and the other with synchronic research. Both parts investigate several aspects of the English language from various perspectives and illustrate the use of corpora in current research. The structure of the volume allows for the same linguistic aspect to be discussed both from the diachronic and the synchronic point of view. The chapters are also useful examples of corpus use as well as of use of other resources as corpus, specifically dictionaries. They investigate a broad array of issues, mainly using corpora of English as a native language, with a focus on corpus tools and corpus description.
This white paper is part of a series that promotes knowledge about language technology and its potential. It addresses educators, journalists, politicians, language communities and others. The availability and use of language technology in Europe varies between languages. Consequently, the actions that are required to further support research and development of language technologies also differ for each language. The required actions depend on many factors, such as the complexity of a given language and the size of its community. META-NET, a Network of Excellence funded by the European Commission, has conducted an analysis of current language resources and technologies. This analysis focused on the 23 official European languages as well as other important national and regional languages in Europe. The results of this analysis suggest that there are many significant research gaps for each language. A more detailed expert analysis and assessment of the current situation will help maximise the impact of additional research and minimize any risks. META-NET consists of 54 research centres from 33 countries that are working with stakeholders from commercial businesses, government agencies, industry, research organisations, software companies, technology providers and European universities. Together, they are creating a common technology vision while developing a strategic research agenda that shows how language technology applications can address any research gaps by 2020.
This book introduces an approach that can be used to ground a variety of intelligent systems, ranging from simple fact based systems to highly sophisticated reasoning systems. As the popularity of AI related fields has grown over the last decade, the number of persons interested in building intelligent systems has increased exponentially. Some of these people are highly skilled and experienced in the use of Al techniques, but many lack that kind of expertise. Much of the literature that might otherwise interest those in the latter category is not appreci ated by them because the material is too technical, often needlessly so. The so called logicists see logic as a primary tool and favor a formal approach to Al, whereas others are more content to rely on informal methods. This polarity has resulted in different styles of writing and reporting, and people entering the field from other disciplines often find themselves hard pressed to keep abreast of current differences in style. This book attempts to strike a balance between these approaches by covering points from both technical and nontechnical perspectives and by doing so in a way that is designed to hold the interest of readers of each persuasion. During recent years, a somewhat overwhelming number of books that present general overviews of Al related subjects have been placed on the market . These books serve an important function by providing researchers and others entering the field with progress reports and new developments.
This white paper is part of a series that promotes knowledge about language technology and its potential. It addresses educators, journalists, politicians, language communities and others. The availability and use of language technology in Europe varies between languages. Consequently, the actions that are required to further support research and development of language technologies also differ for each language. The required actions depend on many factors, such as the complexity of a given language and the size of its community. META-NET, a Network of Excellence funded by the European Commission, has conducted an analysis of current language resources and technologies. This analysis focused on the 23 official European languages as well as other important national and regional languages in Europe. The results of this analysis suggest that there are many significant research gaps for each language. A more detailed expert analysis and assessment of the current situation will help maximise the impact of additional research and minimize any risks. META-NET consists of 54 research centres from 33 countries that are working with stakeholders from commercial businesses, government agencies, industry, research organisations, software companies, technology providers and European universities. Together, they are creating a common technology vision while developing a strategic research agenda that shows how language technology applications can address any research gaps by 2020.
This book constitutes the proceedings of the Third International Conference of the CLEF Initiative, CLEF 2012, held in Rome, Italy, in September 2012. The 14 papers and 3 poster abstracts presented were carefully reviewed and selected for inclusion in this volume. Furthermore, the books contains 2 keynote papers. The papers are organized in topical sections named: benchmarking and evaluation initiatives; information access; and evaluation methodologies and infrastructure.
This white paper is part of a series that promotes knowledge about language technology and its potential. It addresses educators, journalists, politicians, language communities and others. The availability and use of language technology in Europe varies between languages. Consequently, the actions that are required to further support research and development of language technologies also differ for each language. The required actions depend on many factors, such as the complexity of a given language and the size of its community. META-NET, a Network of Excellence funded by the European Commission, has conducted an analysis of current language resources and technologies. This analysis focused on the 23 official European languages as well as other important national and regional languages in Europe. The results of this analysis suggest that there are many significant research gaps for each language. A more detailed expert analysis and assessment of the current situation will help maximise the impact of additional research and minimize any risks. META-NET consists of 54 research centres from 33 countries that are working with stakeholders from commercial businesses, government agencies, industry, research organisations, software companies, technology providers and European universities. Together, they are creating a common technology vision while developing a strategic research agenda that shows how language technology applications can address any research gaps by 2020.
This white paper is part of a series that promotes knowledge about language technology and its potential. It addresses educators, journalists, politicians, language communities and others. The availability and use of language technology in Europe varies between languages. Consequently, the actions that are required to further support research and development of language technologies also differ for each language. The required actions depend on many factors, such as the complexity of a given language and the size of its community. META-NET, a Network of Excellence funded by the European Commission, has conducted an analysis of current language resources and technologies. This analysis focused on the 23 official European languages as well as other important national and regional languages in Europe. The results of this analysis suggest that there are many significant research gaps for each language. A more detailed expert analysis and assessment of the current situation will help maximise the impact of additional research and minimize any risks. META-NET consists of 54 research centres from 33 countries that are working with stakeholders from commercial businesses, government agencies, industry, research organisations, software companies, technology providers and European universities. Together, they are creating a common technology vision while developing a strategic research agenda that shows how language technology applications can address any research gaps by 2020.
"Phonetic Search Methods for Large Databases" focuses on Keyword Spotting (KWS) within large speech databases. The brief will begin by outlining the challenges associated with Keyword Spotting within large speech databases using dynamic keyword vocabularies. It will then continue by highlighting the various market segments in need of KWS solutions, as well as, the specific requirements of each market segment. The work also includes a detailed description of the complexity of the task and the different methods that are used, including the advantages and disadvantages of each method and an in-depth comparison. The main focus will be on the Phonetic Search method and its efficient implementation. This will include a literature review of the various methods used for the efficient implementation of Phonetic Search Keyword Spotting, with an emphasis on the authors' own research which entails a comparative analysis of the Phonetic Search method which includes algorithmic details. This brief is useful for researchers and developers in academia and industry from the fields of speech processing and speech recognition, specifically Keyword Spotting.
Edited in collaboration with FoLLI, the Association of Logic, Language and Information, this book constitutes the refereed proceedings of the 7th International Conference on Logical Aspects of Computational Linguistics, LACL 2012, held in Nantes, France, in July 2012. The 15 revised full papers presented together with 2 invited talks were carefully reviewed and selected from 24 submissions. The papers are organized in topical sections on logical foundation of syntactic formalisms, logics for semantics of lexical items, sentences, discourse and dialog, applications of these models to natural language processing, type theoretic, proof theoretic, model theoretic and other logically based formal methods for describing natural language syntax, semantics and pragmatics, as well as the implementation of natural language processing software relying on such methods.
The contributions to this volume are drawn from the interdisciplinary research c- ried out within the Sonderforschungsbereich (SFB 378), a special long-term funding scheme of the German National Science Foundation (DFG). Sonderforschungsbe- ich 378 was situated at Saarland University, with colleagues from arti?cial intel- gence, computational linguistics, computer science, philosophy, psychology - and in its ?nal phases - cognitive neuroscience and psycholinguistics. The funding covered a period of 12 years, which was split into four phases of 3 years each, ending in December of 2007. Every sub-period culminated in an intensive reviewing process, comprising written reports as well as on-site p- sentations and demonstrations to the external reviewers. We are most grateful to these reviewers for their extensive support and critical feedback; they contributed 1 their time and labor freely to the DFG, the independent and self-organized ins- tution of German scientists. The ?nal evaluation of the DFG reviewers judged the overall performance and the actual work with the highest possible mark, i.e. "excellent".
This book grew out of the Fourth Conference on Computers and the Writing Process, held at the University of Sussex in March 1991. The conference brought together a wide variety of people interested in most aspects of computers and the writing process including, computers and writing education, computer supported fiction, computers and technical writing, evaluation of computer-based writing, and hypertext. Fifteen papers were selected from the twenty-five delivered at the conference. The authors were asked to develop them into articles, incorporating any insights they had gained from their conference presentations. This book offers a survey of the wide area of Computers and Writing, and describes current work in the design and use of computer-based tools for writing. University of Sussex M.S. October, 1991 Note from Publisher This collection of articles is being published simultaneously as a special issue, Volume 21(1-3), of Instructional Science - An International Journal of Learning and Cognition. Instructional Science 21: 1-4 (1992) 1 (c) Kluwer Academic Publishers, Dordrecht Introduction MIKE SHARPLES School of Cognitive and Computing Sciences, University of Sussex, Falmer, Brighton BNl 9QH, United Kingdom.
In opposition to the classical set theory of natural language, Novak's highly original monograph offers a theory based on alternative and fuzzy sets. This new approach is firmly grounded in semantics and pragmatics, and accounts for the vagueness inherent in natural language-filling a large gap in our current knowledge. The theory will foster fruitful debate among researchers in linguistics and artificial intellegence.
In knowledge-based natural language generation, issues of formal knowledge representation meet with the linguistic problems of choosing the most appropriate verbalization in a particular situation of utterance. Lexical Semantics and Knowledge Representation in Multilingual Text Generation presents a new approach to systematically linking the realms of lexical semantics and knowledge represented in a description logic. For language generation from such abstract representations, lexicalization is taken as the central step: when choosing words that cover the various parts of the content representation, the principal decisions on conveying the intended meaning are made. A preference mechanism is used to construct the utterance that is best tailored to parameters representing the context. Lexical Semantics and Knowledge Representation in Multilingual Text Generation develops the means for systematically deriving a set of paraphrases from the same underlying representation with the emphasis on events and verb meaning. Furthermore, the same mapping mechanism is used to achieve multilingual generation: English and German output are produced in parallel, on the basis of an adequate division between language-neutral and language-specific (lexical and grammatical) knowledge. Lexical Semantics and Knowledge Representation in Multilingual Text Generation provides detailed insights into designing the representations and organizing the generation process. Readers with a background in artificial intelligence, cognitive science, knowledge representation, linguistics, or natural language processing will find a model of language production that can be adapted to a variety of purposes.
Derivation or Representation? Hubert Haider & Klaus Netter 1 The Issue Derivation and Representation - these keywords refer both to a conceptual as well as to an empirical issue. Transformational grammar was in its outset (Chomsky 1957, 1975) a derivational theory which characterized a well-formed sentence by its derivation, i.e. a set of syntactic representations defined by a set of rules that map one representation into another. The set of mapping rules, the transformations, eventually became more and more abstract and were trivialized into a single one, namely "move a," a general movement-rule. The constraints on movement were singled out in systems of principles that ap ply to the resulting representations, i.e. the configurations containing a moved element and its extraction site, the trace. The introduction of trace-theory (d. Chomsky 1977, ch.3 17, ch. 4) in principle opened up the possibility of com pletely abandoning movement and generating the possible outputs of movement directly, i.e. as structures that contain gaps representing the extraction sites." |
You may like...
Trends in E-Tools and Resources for…
Gloria Corpas Pastor, Isabel Duran Munoz
Hardcover
R3,025
Discovery Miles 30 250
The Temporal Structure of Multimodal…
Laszlo Hunyadi, Istvan Szekrenyes
Hardcover
R2,653
Discovery Miles 26 530
The Oxford Handbook of Information…
Caroline Fery, Shinichiro Ishihara
Hardcover
R4,569
Discovery Miles 45 690
The Art and Science of Machine…
Walker H. Land Jr., J. David Schaffer
Hardcover
R4,039
Discovery Miles 40 390
The Natural Language for Artificial…
Dioneia Motta Monte-Serrat, Carlo Cattani
Paperback
R2,767
Discovery Miles 27 670
|