![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Language & Literature > Language & linguistics > Computational linguistics
This is the first comprehensive overview of computational approaches to Arabic morphology. The subtitle aims to reflect that widely different computational approaches to the Arabic morphological system have been proposed. The book provides a showcase of the most advanced language technologies applied to one of the most vexing problems in linguistics. It covers knowledge-based and empirical-based approaches.
Many approaches have already been proposed for classification and modeling in the literature. These approaches are usually based on mathematical mod els. Computer systems can easily handle mathematical models even when they are complicated and nonlinear (e.g., neural networks). On the other hand, it is not always easy for human users to intuitively understand mathe matical models even when they are simple and linear. This is because human information processing is based mainly on linguistic knowledge while com puter systems are designed to handle symbolic and numerical information. A large part of our daily communication is based on words. We learn from various media such as books, newspapers, magazines, TV, and the Inter net through words. We also communicate with others through words. While words play a central role in human information processing, linguistic models are not often used in the fields of classification and modeling. If there is no goal other than the maximization of accuracy in classification and model ing, mathematical models may always be preferred to linguistic models. On the other hand, linguistic models may be chosen if emphasis is placed on interpretability."
ThisbookdiscusseshowTypeLogicalGrammarcanbemodi?edinsuch awaythatasystematictreatmentofanaphoraphenomenabecomesp- sible without giving up the general architecture of this framework. By Type Logical Grammar, I mean the version of Categorial Grammar that arose out of the work of Lambek, 1958 and Lambek, 1961. There Ca- gorial types are analyzed as formulae of a logical calculus. In particular, the Categorial slashes are interpreted as forms of constructive impli- tion in the sense of Intuitionistic Logic. Such a theory of grammar is per se attractive for a formal linguist who is interested in the interplay between formal logic and the structure of language. What makes L- bekstyleCategorialGrammarevenmoreexcitingisthefactthat(asvan Benthem,1983pointsout)theCurry-Howardcorrespondence-acentral part of mathematical proof theory which establishes a deep connection betweenconstructivelogicsandthe?-calculus-suppliesthetypelogical syntax with an extremely elegant and independently motivated interface to model-theoretic semantics. Prima facie, anaphora does not 't very well into the Categorial picture of the syntax-semantics interface. The Curry-Howard based composition of meaning operates in a local way, and meaning ass- bly is linear, i.e., every piece of lexical meaning is used exactly once. Anaphora, on the other hand, is in principle unbounded, and it involves by de?nition the multiple use of certain semantic resources. The latter problem has been tackled by several Categorial grammarians by ass- ing su?ciently complex lexical meanings for anaphoric expressions, but the locality problem is not easy to solve in a purely lexical way.
This book constitutes the refereed proceedings of the 4th Language and Technology Conference: Challenges for Computer Science and Linguistics, LTC 2009, held in Poznan, Poland, in November 2009. The 52 revised and in many cases substantially extended papers presented in this volume were carefully reviewed and selected from 103 submissions. The contributions are organized in topical sections on speech processing, computational morphology/lexicography, parsing, computational semantics, dialogue modeling and processing, digital language resources, WordNet, document processing, information processing, and machine translation.
This two-volume set, consisting of LNCS 6608 and LNCS 6609, constitutes the thoroughly refereed proceedings of the 12th International Conference on Computer Linguistics and Intelligent Processing, held in Tokyo, Japan, in February 2011. The 74 full papers, presented together with 4 invited papers, were carefully reviewed and selected from 298 submissions. The contents have been ordered according to the following topical sections: lexical resources; syntax and parsing; part-of-speech tagging and morphology; word sense disambiguation; semantics and discourse; opinion mining and sentiment detection; text generation; machine translation and multilingualism; information extraction and information retrieval; text categorization and classification; summarization and recognizing textual entailment; authoring aid, error correction, and style analysis; and speech recognition and generation.
The ninth campaign of the Cross-Language Evaluation Forum (CLEF) for European languages was held from January to September 2008. There were seven main eval- tion tracks in CLEF 2008 plus two pilot tasks. The aim, as usual, was to test the p- formance of a wide range of multilingual information access (MLIA) systems or s- tem components. This year, 100 groups, mainly but not only from academia, parti- pated in the campaign. Most of the groups were from Europe but there was also a good contingent from North America and Asia plus a few participants from South America and Africa. Full details regarding the design of the tracks, the methodologies used for evaluation, and the results obtained by the participants can be found in the different sections of these proceedings. The results of the CLEF 2008 campaign were presented at a two-and-a-half day workshop held in Aarhus, Denmark, September 17-19, and attended by 150 resear- ers and system developers. The annual workshop, held in conjunction with the European Conference on Digital Libraries, plays an important role by providing the opportunity for all the groups that have participated in the evaluation campaign to get together comparing approaches and exchanging ideas. The schedule of the workshop was divided between plenary track overviews, and parallel, poster and breakout sessions presenting this year's experiments and discu- ing ideas for the future. There were several invited talks.
Introduces the two most common numerical methods for heat transfer and fluid dynamics equations, using clear and accessible language. This unique approach covers all necessary mathematical preliminaries at the beginning of the book for the reader to sail smoothly through the chapters. Students will work step-by-step through the most common benchmark heat transfer and fluid dynamics problems, firmly grounding themselves in how the governing equations are discretized, how boundary conditions are imposed, and how the resulting algebraic equations are solved. Providing a detailed discussion of the discretization steps and time approximations, and clearly presenting concepts of explicit and implicit formulations, this graduate textbook has everything an instructor needs to prepare students for their exams and future careers. Each illustrative example shows students how to draw comparisons between the results obtained using the two numerical methods, and at the end of each chapter they can test and extend their understanding by working through the problems provided. A solutions manual is also available for instructors.
th TSD 2009was the 12 eventin the series of InternationalConferenceson Text, Speech andDialoguesupportedbytheInternationalSpeechCommunicationAssociation(ISCA) ? and Czech Society for Cybernetics and Informatics (CSKI). This year, TSD was held in Plzen ? (Pilsen), in the Primavera Conference Center, during September 13-17, 2009 and it was organized by the University of West Bohemia in Plzen ? in cooperation with Masaryk University of Brno, Czech Republic. Like its predecessors, TSD 2009 hi- lighted to both the academic and scienti?c world the importance of text and speech processing and its most recent breakthroughsin current applications. Both experienced researchers and professionals as well as newcomers to the text and speech processing ?eld, interested in designing or evaluating interactive software, developing new int- action technologies, or investigatingoverarchingtheories of text and speech processing found in the TSD conference a forum to communicate with people sharing similar - terests. The conference is an interdisciplinary forum, intertwining research in speech and language processing with its applications in everyday practice. We feel that the mixture of different approaches and applications offered a great opportunity to get - quaintedwith currentactivitiesin all aspects oflanguagecommunicationand to witness the amazing vitality of researchers from developing countries too. This year's conference was partially oriented toward semantic processing, which was chosen as the main topic of the conference. All invited speakers (Frederick Jelinek, Louise Guthrie, Roberto Pieraccini, Tilman Becker, and Elmar Not ] h) gave lectures on thenewestresultsintherelativelybroadandstillunexploredareaofsemanticprocessing."
Marcus Contextual Grammars is the first monograph to present a class of grammars introduced about three decades ago, based on the fundamental linguistic phenomenon of strings-contexts interplay (selection). Most of the theoretical results obtained so far about the many variants of contextual grammars are presented with emphasis on classes of questions with relevance for applications in the study of natural language syntax: generative powers, descriptive and computational complexity, automata recognition, semilinearity, structure of the generated strings, ambiguity, regulated rewriting, etc. Constant comparison with families of languages in the Chomsky hierarchy is made. Connections with non-linguistic areas are established, such as molecular computing. Audience: Researchers and students in theoretical computer science (formal language theory and automata theory), computational linguistics, mathematical methods in linguistics, and linguists interested in formal models of syntax.
This book offers a state-of-the-art survey of methods and techniques for structuring, acquiring and maintaining lexical resources for speech and language processing. The first chapter provides a broad survey of the field of computational lexicography, introducing most of the issues, terms and topics which are addressed in more detail in the rest of the book. The next two chapters focus on the structure and the content of man-made lexicons, concentrating respectively on (morpho-)syntactic and (morpho-)phonological information. Both chapters adopt a declarative constraint-based methodology and pay ample attention to the various ways in which lexical generalizations can be formalized and exploited to enhance the consistency and to reduce the redundancy of lexicons. A complementary perspective is offered in the next two chapters, which present techniques for automatically deriving lexical resources from text corpora. These chapters adopt an inductive data-oriented methodology and focus also on methods for tokenization, lemmatization and shallow parsing. The next three chapters focus on speech applications, more specifically on the organization of speech data bases, and on the use of lexica in speech synthesis and speech recognition. The last chapter takes a psycholinguistic perspective and addresses the relation between storage and computation in the mental lexicon. The relevance of these topics for speech and language processing is obvious, for since NLP systems need large lexica in order to achieve reasonable coverage, and since the construction and maintenance of large-size lexical resources is a complex and costly task, it is of crucial importance for those who design or build such systems to be aware of the latest developments in this fast-moving field. The intended audience for this book includes advanced students and professional scientists working in the areas of computational linguistics and language and speech technology.
This book is based on contributions to the Seventh European Summer School on Language and Speech Communication that was held at KTH in Stockholm, Sweden, in July of 1999 under the auspices of the European Language and Speech Network (ELSNET). The topic of the summer school was "Multimodality in Language and Speech Systems" (MiLaSS). The issue of multimodality in interpersonal, face-to-face communication has been an important research topic for a number of years. With the increasing sophistication of computer-based interactive systems using language and speech, the topic of multimodal interaction has received renewed interest both in terms of human-human interaction and human-machine interaction. Nine lecturers contri buted to the summer school with courses on specialized topics ranging from the technology and science of creating talking faces to human-human communication, which is mediated by computer for the handicapped. Eight of the nine lecturers are represented in this book. The summer school attracted more than 60 participants from Europe, Asia and North America representing not only graduate students but also senior researchers from both academia and industry."
The present volume contributes to the growing body of work on sentence pro- cessing. The goal of work in this area is to construct a theory of human sen- tence processing in general, i.e., given a grammar of some particular language and a general characterization of the human sentence processing mechanisms, the particular processing system for that language should follows automati- cally. At least that's the goal. What is needed in order to pursue this goal is systematic in-depth analysis of the sentence routines of individual languages. With respect to German, that is precisely what the present volume delivers. In sharp contrast to a decade ago, the study of German sentence process- ing is flourishing today. Four lively and active centers have emerged. The University of Freiburg is one prominent center, represented in the present vol- ume by the editors Barbara Hemforth and Lars Konieczny (who was at Freiburg for many years) as well as by Christoph Scheepers (who is now in Glasgow) and Christoph Holscher. The University of Potsdam has recently begun an interdisciplinary collaboration on sentence processing involving Matthias Schlesewsky, Gisbert Fanselow, Reinhold Kliegl and Josef Krems. The University of Jena has several investigators trained in linguistics and interested in language processing. That group is represented here by Markus Bader and also includes his colleagues Michael Meng and Josef Bayer.
The eleven chapters of this book represent an original contribution to the field of multimodal spoken dialogue systems. The material includes highly relevant topics, such as dialogue modeling in research systems versus industrial systems. The book contains detailed application studies, including speech-controlled MP3 players in a car environment, negotiation training with a virtual human in a military context and the application of spoken dialogue to question-answering systems.
Automatic Text Categorization and Clustering are becoming more and more important as the amount of text in electronic format grows and the access to it becomes more necessary and widespread. Well known applications are spam filtering and web search, but a large number of everyday uses exist (intelligent web search, data mining, law enforcement, etc.) Currently, researchers are employing many intelligent techniques for text categorization and clustering, ranging from support vector machines and neural networks to Bayesian inference and algebraic methods, such as Latent Semantic Indexing. This volume offers a wide spectrum of research work developed for intelligent text categorization and clustering. In the following, we give a brief introduction of the chapters that are included in this book.
In both the linguistic and the language engineering community, the creation and use of annotated text collections (or annotated corpora) is currently a hot topic. Annotated texts are of interest for research as well as for the development of natural language pro cessing (NLP) applications. Unfortunately, the annotation of text material, especially more interesting linguistic annotation, is as yet a difficult task and can entail a substan tial amount of human involvement. Allover the world, work is being done to replace as much as possible of this human effort by computer processing. At the frontier of what can already be done (mostly) automatically we find syntactic wordclass tagging, the annotation of the individual words in a text with an indication of their morpho syntactic classification. This book describes the state of the art in syntactic wordclass tagging. As an attempt to give an overall view of the field, this book is of interest to (at least) two, possibly very different, types of reader. The first type consists of those people who are using, or are planning to use, tagged material and taggers. They will want to know what the possibilities and impossibilities of tagging are, but are not necessarily interested in the internal working of automatic taggers. This, on the other hand, is the main interest of our second type of reader, the builders of automatic taggers and other natural language processing software."
Natural Language Processing and Text Mining not only discusses applications of Natural Language Processing techniques to certain Text Mining tasks, but also the converse, the use of Text Mining to assist NLP. It assembles a diverse views from internationally recognized researchers and emphasizes caveats in the attempt to apply Natural Language Processing to text mining. This state-of-the-art survey is a must-have for advanced students, professionals, and researchers.
Computational semantics is the art and science of computing meaning in natural language. The meaning of a sentence is derived from the meanings of the individual words in it, and this process can be made so precise that it can be implemented on a computer. Designed for students of linguistics, computer science, logic and philosophy, this comprehensive text shows how to compute meaning using the functional programming language Haskell. It deals with both denotational meaning (where meaning comes from knowing the conditions of truth in situations), and operational meaning (where meaning is an instruction for performing cognitive action). Including a discussion of recent developments in logic, it will be invaluable to linguistics students wanting to apply logic to their studies, logic students wishing to learn how their subject can be applied to linguistics, and functional programmers interested in natural language processing as a new application area.
This book presents a critical overview of current work on
linguistic features and establishes new bases for their use in the
study and understanding of language.
This book teaches the principles of natural language processing and covers linguistics issues. It also details the language-processing functions involved, including part-of-speech tagging using rules and stochastic techniques. A key feature of the book is the author's hands-on approach throughout, with extensive exercises, sample code in Prolog and Perl, and a detailed introduction to Prolog. The book is suitable for researchers and students of natural language processing and computational linguistics.
The relation between ontologies and language is currently at the forefront of natural language processing (NLP). Ontologies, as widely used models in semantic technologies, have much in common with the lexicon. A lexicon organizes words as a conventional inventory of concepts, while an ontology formalizes concepts and their logical relations. A shared lexicon is the prerequisite for knowledge-sharing through language, and a shared ontology is the prerequisite for knowledge-sharing through information technology. In building models of language, computational linguists must be able to accurately map the relations between words and the concepts that they can be linked to. This book focuses on the technology involved in enabling integration between lexical resources and semantic technologies. It will be of interest to researchers and graduate students in NLP, computational linguistics, and knowledge engineering, as well as in semantics, psycholinguistics, lexicology and morphology/syntax.
This volume presents the proceedings of the Third International Sanskrit C- putational Linguistics Symposium hosted by the University of Hyderabad, Hyderabad, IndiaduringJanuary15-17,2009.TheseriesofsymposiaonSanskrit Computational Linguistics began in 2007. The ?rst symposium was hosted by INRIA atRocquencourt, Francein October 2007asa partofthe jointcollabo- tion between INRIA and the University of Hyderabad. This joint collaboration expanded both geographically as well as academically covering more facets of Sanskrit Computaional Linguistics, when the second symposium was hosted by Brown University, USA in May 2008. We received 16 submissions, which were reviewed by the members of the Program Committee. After discussion, nine of them were selected for presen- tion. These nine papers fall under four broad categories: four papers deal with the structure of Pan - ini's Astad - hyay - - ?. Two of them deal with parsing issues, . .. two with various aspects of machine translation, and the last one with the Web concordance of an important Sanskrit text. Ifwelookretrospectivelyoverthelasttwoyears, thethreesymposiainsucc- sion have seen not only continuity of some of the themes, but also steady growth of the community. As is evident, researchers from diverse disciplines such as l- guistics, computer science, philology, and vy- akarana are collaborating with the . scholars from other disciplines, witnessing the growth of Sanskrit computational linguistics as an emergent discipline. We are grateful to S.D. Joshi, Jan Houben, and K.V.R. Krishnamacharyulu for accepting our invitation to deliver the invited speeches."
The International Conference on Computational Processing on Portuguese, f- merly the Workshop on Computational Processing of the Portuguese Language - PROPOR- is the main event in the area of Natural LanguageProcessing that focusesonPortugueseandthe theoreticalandtechnologicalissuesrelatedto this speci?c language. The meeting has been a very rich forum for the interchange of ideas and partnerships for the researchcommunities dedicated to the automated processing of the Portuguese language. This year's PROPOR, the ?rst one to adopt the International Conference - bel, followedworkshopsheld in Lisbon, Portugal(1993), Curitiba, Brazil(1996), PortoAlegre, Brazil(1998), Evora, Portugal(1999), Atibaia, Brazil(2000), Faro, Portugal (2003) and Itatiaia, Brazil (2006). The constitutionofasteeringcommittee (PROPORCommittee), aninter- tional program committee, the adoption of high-standard refereing procedures and the support of the prestigious ACL and ISCA international associations demonstrate the steady development of the ?eld and of its scienti?c community. A total of 63 papers were submitted to PROPOR 2008. Each submitted paper received a careful, triple-blind review by the program committee or by their commitment. All those who contributed are mentioned on the following pages. The reviewing process led to the selection of 21 regular papers for oral presentation and 16 short papers for poster sessions. The workshop and this book were structured around the following main t- ics: Speech Analysis; Ontologies, Semantics and Anaphora Resolution; Speech Synthesis; Machine Learning Applied to Natural Language Processing; Speech Recognition and Natural Language Processing Tools and Applications. Short papers and related posters were organized according to the two main areas of PROPOR: Natural Language Processing and Speech Technology.
In its nine chapters, this book provides an overview of the state-of-the-art and best practice in several sub-fields of evaluation of text and speech systems and components. The evaluation aspects covered include speech and speaker recognition, speech synthesis, animated talking agents, part-of-speech tagging, parsing, and natural language software like machine translation, information retrieval, question answering, spoken dialogue systems, data resources, and annotation schemes. With its broad coverage and original contributions this book is unique in the field of evaluation of speech and language technology. This book is of particular relevance to advanced undergraduate students, PhD students, academic and industrial researchers, and practitioners.
This volume constitutes the thoroughly refereed post-conference proceedings of the First and Second International Symposia on Sanskrit Computational Linguistics, held in Rocquencourt, France, in October 2007 and in Providence, RI, USA, in May 2008 respectively. The 11 revised full papers of the first and the 12 revised papers of the second symposium presented with an introduction and a keynote talk were carefully reviewed and selected from the lectures given at both events. The papers address several topics such as the structure of the Paninian grammatical system, computational linguistics, lexicography, lexical databases, formal description of sanskrit grammar, phonology and morphology, machine translation, philology, and OCR.
The book provides an overview of more than a decade of joint R&D efforts in the Low Countries on HLT for Dutch. It not only presents the state of the art of HLT for Dutch in the areas covered, but, even more importantly, a description of the resources (data and tools) for Dutch that have been created are now available for both academia and industry worldwide. The contributions cover many areas of human language technology (for Dutch): corpus collection (including IPR issues) and building (in particular one corpus aiming at a collection of 500M word tokens), lexicology, anaphora resolution, a semantic network, parsing technology, speech recognition, machine translation, text (summaries) generation, web mining, information extraction, and text to speech to name the most important ones. The book also shows how a medium-sized language community (spanning two territories) can create a digital language infrastructure (resources, tools, etc.) as a basis for subsequent R&D. At the same time, it bundles contributions of almost all the HLT research groups in Flanders and the Netherlands, hence offers a view of their recent research activities. Targeted readers are mainly researchers in human language technology, in particular those focusing on Dutch. It concerns researchers active in larger networks such as the CLARIN, META-NET, FLaReNet and participating in conferences such as ACL, EACL, NAACL, COLING, RANLP, CICling, LREC, CLIN and DIR ( both in the Low Countries), InterSpeech, ASRU, ICASSP, ISCA, EUSIPCO, CLEF, TREC, etc. In addition, some chapters are interesting for human language technology policy makers and even for science policy makers in general. " |
You may like...
The Oxford Handbook of Information…
Caroline Fery, Shinichiro Ishihara
Hardcover
R4,569
Discovery Miles 45 690
Artificial Intelligence for Healthcare…
Boris Galitsky, Saveli Goldberg
Paperback
R2,991
Discovery Miles 29 910
The Natural Language for Artificial…
Dioneia Motta Monte-Serrat, Carlo Cattani
Paperback
R2,767
Discovery Miles 27 670
Machine Learning for Biometrics…
Partha Pratim Sarangi, Madhumita Panda, …
Paperback
R2,570
Discovery Miles 25 700
Spelling and Writing Words - Theoretical…
Cyril Perret, Thierry Olive
Hardcover
R2,803
Discovery Miles 28 030
|