![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Artificial intelligence > Natural language & machine translation
This book constitutes the refereed proceedings of the 16th International Conference on Asia-Pacific Digital Libraries, ICADL 2014, held in Chiang Mai, Thailand, in November 2014. The 20 full papers, 19 short papers and 9 poster papers presented were carefully reviewed and selected from 141 submissions. The papers are organized in topical sections on digital preservation and archiving; digital repositories and tools; scholarly documents repositories; metadata and ontologies; linked data and knowledge sharing; digital books and e-books; digital libraries usage and applications; data management and classification; information retrieval and search methods; user skills and experiences.
This book constitutes the proceedings of the Third International Conference on Analysis of Images, Social Networks and Texts, AIST 2014, held in Yekaterinburg, Russia, in April 2014. The 11 full and 10 short papers were carefully reviewed and selected from 74 submissions. They are presented together with 3 short industrial papers, 4 invited papers and tutorials. The papers deal with topics such as analysis of images and videos; natural language processing and computational linguistics; social network analysis; machine learning and data mining; recommender systems and collaborative technologies; semantic web, ontologies and their applications; analysis of socio-economic data.
This Festschrift volume is published in Honor of Yaacov Choueka on
the occasion of this 75th birthday. The present three-volumes liber
amicorum, several years in gestation, honours this outstanding
Israeli computer scientist and is dedicated to him and to his
scientific endeavours. Yaacov's research has had a major impact not
only within the walls of academia, but also in the daily life of
lay users of such technology that originated from his research. An
especially amazing aspect of the temporal span of his scholarly
work is that half a century after his influential research from the
early 1960s, a project in which he is currently involved is proving
to be a sensation, as will become apparent from what follows.
Yaacov Choueka began his research career in the theory of computer
science, dealing with basic questions regarding the relation
between mathematical logic and automata theory. From formal
languages, Yaacov moved to natural languages. He was a founder of
natural-language processing in Israel, developing numerous tools
for Hebrew. He is best known for his primary role, together with
Aviezri Fraenkel, in the development of the Responsa Project, one
of the earliest fulltext retrieval systems in the world. More
recently, he has headed the Friedberg Genizah Project, which is
bringing the treasures of the Cairo Genizah into the Digital
Age.
Learning to rank refers to machine learning techniques for training a model in a ranking task. Learning to rank is useful for many applications in information retrieval, natural language processing, and data mining. Intensive studies have been conducted on its problems recently, and significant progress has been made. This lecture gives an introduction to the area including the fundamental problems, major approaches, theories, applications, and future work. The author begins by showing that various ranking problems in information retrieval and natural language processing can be formalized as two basic ranking tasks, namely ranking creation (or simply ranking) and ranking aggregation. In ranking creation, given a request, one wants to generate a ranking list of offerings based on the features derived from the request and the offerings. In ranking aggregation, given a request, as well as a number of ranking lists of offerings, one wants to generate a new ranking list of the offerings. Ranking creation (or ranking) is the major problem in learning to rank. It is usually formalized as a supervised learning task. The author gives detailed explanations on learning for ranking creation and ranking aggregation, including training and testing, evaluation, feature creation, and major approaches. Many methods have been proposed for ranking creation. The methods can be categorized as the pointwise, pairwise, and listwise approaches according to the loss functions they employ. They can also be categorized according to the techniques they employ, such as the SVM based, Boosting based, and Neural Network based approaches. The author also introduces some popular learning to rank methods in details. These include: PRank, OC SVM, McRank, Ranking SVM, IR SVM, GBRank, RankNet, ListNet & ListMLE, AdaRank, SVM MAP, SoftRank, LambdaRank, LambdaMART, Borda Count, Markov Chain, and CRanking. The author explains several example applications of learning to rank including web search, collaborative filtering, definition search, keyphrase extraction, query dependent summarization, and re-ranking in machine translation. A formulation of learning for ranking creation is given in the statistical learning framework. Ongoing and future research directions for learning to rank are also discussed. Table of Contents: Learning to Rank / Learning for Ranking Creation / Learning for Ranking Aggregation / Methods of Learning to Rank / Applications of Learning to Rank / Theory of Learning to Rank / Ongoing and Future Work
This Festschrift volume is published in Honor of Yaacov Choueka on the occasion of this 75th birthday. The present three-volumes liber amicorum, several years in gestation, honours this outstanding Israeli computer scientist and is dedicated to him and to his scientific endeavours. Yaacov's research has had a major impact not only within the walls of academia, but also in the daily life of lay users of such technology that originated from his research. An especially amazing aspect of the temporal span of his scholarly work is that half a century after his influential research from the early 1960s, a project in which he is currently involved is proving to be a sensation, as will become apparent from what follows. Yaacov Choueka began his research career in the theory of computer science, dealing with basic questions regarding the relation between mathematical logic and automata theory. From formal languages, Yaacov moved to natural languages. He was a founder of natural-language processing in Israel, developing numerous tools for Hebrew. He is best known for his primary role, together with Aviezri Fraenkel, in the development of the Responsa Project, one of the earliest fulltext retrieval systems in the world. More recently, he has headed the Friedberg Genizah Project, which is bringing the treasures of the Cairo Genizah into the Digital Age. This third part of the three-volume set covers a range of topics related to language, ranging from linguistics to applications of computation to language, using linguistic tools. The papers are grouped in topical sections on: natural language processing; representing the lexicon; and neologisation.
This Festschrift volume is published in Honor of Yaacov Choueka on the occasion of this 75th birthday. The present three-volumes liber amicorum, several years in gestation, honours this outstanding Israeli computer scientist and is dedicated to him and to his scientific endeavours. Yaacov's research has had a major impact not only within the walls of academia, but also in the daily life of lay users of such technology that originated from his research. An especially amazing aspect of the temporal span of his scholarly work is that half a century after his influential research from the early 1960s, a project in which he is currently involved is proving to be a sensation, as will become apparent from what follows. Yaacov Choueka began his research career in the theory of computer science, dealing with basic questions regarding the relation between mathematical logic and automata theory. From formal languages, Yaacov moved to natural languages. He was a founder of natural-language processing in Israel, developing numerous tools for Hebrew. He is best known for his primary role, together with Aviezri Fraenkel, in the development of the Responsa Project, one of the earliest fulltext retrieval systems in the world. More recently, he has headed the Friedberg Genizah Project, which is bringing the treasures of the Cairo Genizah into the Digital Age. This first part of the three-volume set covers a range of topics in computer science. The papers are grouped in topical sections on: the jubilaris: Yaacov and his oeuvre; theory of computation; science computing and tools for engineering; information retrieval.
This book constitutes the refereed proceedings of the 5th International Conference of the CLEF Initiative, CLEF 2014, held in Sheffield, UK, in September 2014. The 11 full papers and 5 short papers presented were carefully reviewed and selected from 30 submissions. They cover a broad range of issues in the fields of multilingual and multimodal information access evaluation, also included are a set of labs and workshops designed to test different aspects of mono and cross-language information retrieval systems
This book constitutes the joint refereed proceedings of Calculemus 2014, Digital Mathematics Libraries, DML 2014, Mathematical Knowledge Management, MKM 2014 and Systems and Projects, S&P 2014, held in Coimbra, Portugal, during July 7-11, 2014 as four tracks of CICM 2014, the Conferences on Intelligent Computer Mathematics. The 26 full papers and 9 Systems and Projects descriptions presented together with 5 invited talks were carefully reviewed and selected from a total of 55 submissions. The Calculemus track of CICM examines the integration of symbolic computation and mechanized reasoning. The Digital Mathematics Libraries track - evolved from the DML workshop series - features math-aware technologies, standards, algorithms and processes towards the fulfillment of the dream of a global DML. The Mathematical Knowledge Management track of CICM is concerned with all aspects of managing mathematical knowledge in the informal, semi-formal and formal settings. The Systems and Projects track presents short descriptions of existing systems or on-going projects in the areas of all the other tracks of the conference.
Edited in collaboration with FoLLI, the Association of Logic, Language and Information, this book constitutes the refereed proceedings of the 8th International Conference on Logical Aspects of Computational Linguistics (LACL 2014) held in Toulouse, France, in June 2014. On the broadly syntactic side, there are papers on the logical and computational foundations of context free grammars, pregroup grammars, on the Lambek calculus and on formalizations of aspects of minimalism. There is also a paper on Abstract Categorical Grammar, as well as papers on issues at the syntax/semantics interface. On the semantic side, the volume's papers address monotonicity reasoning and the semantics of adverbs in type theory, proof theoretical semantics and predicate and argument invariance.
This book constitutes the refereed proceedings of the 12th International Conference on Intelligent Tutoring Systems, ITS 2014, held in Honolulu, HI, USA, in June 2014. The 31 revised full papers, 45 short papers and 27 posters presented were carefully viewed and selected from 177 submissions. The specific theme of the ITS 2014 conference is "Creating fertile soil for learning interactions." Besides that, the highly interdisciplinary ITS conferences bring together researchers in computer science, learning sciences, cognitive and educational psychology, sociology, cognitive science, artificial intelligence, machine learning and linguistics. The papers are organized in topical sections on affect; multimodality and metacognition; collaborative learning; data mining and student behavior; dialogue and discourse; generating hints, scaffolds and questions; game-based learning and simulation; graphical representations and learning; student strategies and problem solving; scaling ITS and assessment.
Written by leading international experts, this volume presents contributions establishing the feasibility of human language-like communication with robots. The book explores the use of language games for structuring situated dialogues in which contextualized language communication and language acquisition can take place. Within the text are integrated experiments demonstrating the extensive research which targets artificial language evolution. Language Grounding in Robots uses the design layers necessary to create a fully operational communicating robot as a framework for the text, focusing on the following areas: Embodiment; Behavior; Perception and Action; Conceptualization; Language Processing; Whole Systems Experiments. This book serves as an excellent reference for researchers interested in further study of artificial language evolution."
This book constitutes the thoroughly refereed post-workshop proceedings of the 5th International Workshop on Camera-Based Document Analysis and Recognition, CBDAR 2013, held in Washington, DC, USA, in August 2013. The 14 revised full papers presented were carefully selected during two rounds of reviewing and improvement from numerous original submissions. Intended to give a snapshot of the state-of-the-art research in the field of camera based document analysis and recognition, the papers are organized in topical sections on text detection and recognition in scene images and camera-based systems.
For humans, understanding a natural language sentence or discourse is so effortless that we hardly ever think about it. For machines, however, the task of interpreting natural language, especially grasping meaning beyond the literal content, has proven extremely difficult and requires a large amount of background knowledge. This book focuses on the interpretation of natural language with respect to specific domain knowledge captured in ontologies. The main contribution is an approach that puts ontologies at the center of the interpretation process. This means that ontologies not only provide a formalization of domain knowledge necessary for interpretation but also support and guide the construction of meaning representations. We start with an introduction to ontologies and demonstrate how linguistic information can be attached to them by means of the ontology lexicon model lemon. These lexica then serve as basis for the automatic generation of grammars, which we use to compositionally construct meaning representations that conform with the vocabulary of an underlying ontology. As a result, the level of representational granularity is not driven by language but by the semantic distinctions made in the underlying ontology and thus by distinctions that are relevant in the context of a particular domain. We highlight some of the challenges involved in the construction of ontology-based meaning representations, and show how ontologies can be exploited for ambiguity resolution and the interpretation of temporal expressions. Finally, we present a question answering system that combines all tools and techniques introduced throughout the book in a real-world application, and sketch how the presented approach can scale to larger, multi-domain scenarios in the context of the Semantic Web. Table of Contents: List of Figures / Preface / Acknowledgments / Introduction / Ontologies / Linguistic Formalisms / Ontology Lexica / Grammar Generation / Putting Everything Together / Ontological Reasoning for Ambiguity Resolution / Temporal Interpretation / Ontology-Based Interpretation for Question Answering / Conclusion / Bibliography / Authors' Biographies
It has been estimated that over a billion people are using or learning English as a second or foreign language, and the numbers are growing not only for English but for other languages as well. These language learners provide a burgeoning market for tools that help identify and correct learners' writing errors. Unfortunately, the errors targeted by typical commercial proofreading tools do not include those aspects of a second language that are hardest to learn. This volume describes the types of constructions English language learners find most difficult: constructions containing prepositions, articles, and collocations. It provides an overview of the automated approaches that have been developed to identify and correct these and other classes of learner errors in a number of languages. Error annotation and system evaluation are particularly important topics in grammatical error detection because there are no commonly accepted standards. Chapters in the book describe the options available to researchers, recommend best practices for reporting results, and present annotation and evaluation schemes. The final chapters explore recent innovative work that opens new directions for research. It is the authors' hope that this volume will continue to contribute to the growing interest in grammatical error detection by encouraging researchers to take a closer look at the field and its many challenging problems.
This white paper is part of a series that promotes knowledge about language technology and its potential. It addresses educators, journalists, politicians, language communities and others. The availability and use of language technology in Europe varies between languages. Consequently, the actions that are required to further support research and development of language technologies also differ for each language. The required actions depend on many factors, such as the complexity of a given language and the size of its community. META-NET, a Network of Excellence funded by the European Commission, has conducted an analysis of current language resources and technologies. This analysis focused on the 23 official European languages as well as other important national and regional languages in Europe. The results of this analysis suggest that there are many significant research gaps for each language. A more detailed expert analysis and assessment of the current situation will help maximise the impact of additional research and minimize any risks. META-NET consists of 54 research centres from 33 countries that are working with stakeholders from commercial businesses, government agencies, industry, research organisations, software companies, technology providers and European universities. Together, they are creating a common technology vision while developing a strategic research agenda that shows how language technology applications can address any research gaps by 2020.
Natural language is one of the most important means of human communication. It enables us to express our will, to exchange thoughts and to document our knowledge in written sources. Owing to its substantial role in many facets of human life, technology for automatically analyzing and processing natural language has recently become increasingly important. In fact, natural language processing tools have paved the way for entirely new business opportunities. The goal of this book is to facilitate the automatic analysis of natural language in process models and to employ this analysis for assisting process model stakeholders. Therefore, a technique is defined that automatically recognizes and annotates process model element labels. In addition, this technique is leveraged to support organizations in effectively utilizing their process models in various ways. The book is organized into seven chapters. It starts with an overview of business process management and linguistics and continues with conceptual contributions on parsing and annotating process model elements, with the detection and correction of process model guideline violations, with the generation of natural language from process models and finally ends with the derivation of service candidates from process models.
This book constitutes the refereed proceedings of the International Conference on Theory and Practice of Digital Libraries, TPDL 2013 (formerly European Conference on Research and Advanced Technology for Digital Libraries, ECDL) held in Valletta, Malta, in September 2013. The 24 full papers, 13 short papers, 22 posters and 8 demonstrations presented in this volume were carefully reviewed and selected from 158 submissions. The papers cover a wide range of research topics, clustered in four broader areas: foundation, infrastructures, content, and services. They have been organized in topical sections on conceptual models and formal issues, aggregation and archiving, user behavior, digital curation, mining and extraction, architectures and interoperability, interfaces to digital libraries, semantic web, information retrieval and browsing, and preservation. Also included are 6 tutorials and 2 panels.
The World Wide Web constitutes the largest existing source of texts written in a great variety of languages. A feasible and sound way of exploiting this data for linguistic research is to compile a static corpus for a given language. There are several adavantages of this approach: (i) Working with such corpora obviates the problems encountered when using Internet search engines in quantitative linguistic research (such as non-transparent ranking algorithms). (ii) Creating a corpus from web data is virtually free. (iii) The size of corpora compiled from the WWW may exceed by several orders of magnitudes the size of language resources offered elsewhere. (iv) The data is locally available to the user, and it can be linguistically post-processed and queried with the tools preferred by her/him. This book addresses the main practical tasks in the creation of web corpora up to giga-token size. Among these tasks are the sampling process (i.e., web crawling) and the usual cleanups including boilerplate removal and removal of duplicated content. Linguistic processing and problems with linguistic processing coming from the different kinds of noise in web corpora are also covered. Finally, the authors show how web corpora can be evaluated and compared to other corpora (such as traditionally compiled corpora). For additional material please visit the companion website: sites.morganclaypool.com/wcc Table of Contents: Preface / Acknowledgments / Web Corpora / Data Collection / Post-Processing / Linguistic Processing / Corpus Evaluation and Comparison / Bibliography / Authors' Biographies
Design and create neural networks with deep learning and artificial intelligence principles using OpenAI Gym, TensorFlow, and Keras Key Features Explore neural network architecture and understand how it functions Learn algorithms to solve common problems using back propagation and perceptrons Understand how to apply neural networks to applications with the help of useful illustrations Book DescriptionNeural networks play a very important role in deep learning and artificial intelligence (AI), with applications in a wide variety of domains, right from medical diagnosis, to financial forecasting, and even machine diagnostics. Hands-On Neural Networks is designed to guide you through learning about neural networks in a practical way. The book will get you started by giving you a brief introduction to perceptron networks. You will then gain insights into machine learning and also understand what the future of AI could look like. Next, you will study how embeddings can be used to process textual data and the role of long short-term memory networks (LSTMs) in helping you solve common natural language processing (NLP) problems. The later chapters will demonstrate how you can implement advanced concepts including transfer learning, generative adversarial networks (GANs), autoencoders, and reinforcement learning. Finally, you can look forward to further content on the latest advancements in the field of neural networks. By the end of this book, you will have the skills you need to build, train, and optimize your own neural network model that can be used to provide predictable solutions. What you will learn Learn how to train a network by using backpropagation Discover how to load and transform images for use in neural networks Study how neural networks can be applied to a varied set of applications Solve common challenges faced in neural network development Understand the transfer learning concept to solve tasks using Keras and Visual Geometry Group (VGG) network Get up to speed with advanced and complex deep learning concepts like LSTMs and NLP Explore innovative algorithms like GANs and deep reinforcement learning Who this book is forIf you are interested in artificial intelligence and deep learning and want to further your skills, then this intermediate-level book is for you. Some knowledge of statistics will help you get the most out of this book.
This book evaluates the impact of relevant factors affecting the results of speech quality assessment studies carried out in crowdsourcing. The author describes how these factors relate to the test structure, the effect of environmental background noise, and the influence of language differences. He details multiple user-centered studies that have been conducted to derive guidelines for reliable collection of speech quality scores in crowdsourcing. Specifically, different questions are addressed such as the optimal number of speech samples to include in a listening task, the influence of the environmental background noise in the speech quality ratings, as well as methods for classifying background noise from web audio recordings, or the impact of language proficiency in the user perception of speech quality. Ultimately, the results of these studies contributed to the definition of the ITU-T Recommendation P.808 that defines the guidelines to conduct speech quality studies in crowdsourcing.
In the last few years, a number of NLP researchers have developed and participated in the task of Recognizing Textual Entailment (RTE). This task encapsulates Natural Language Understanding capabilities within a very simple interface: recognizing when the meaning of a text snippet is contained in the meaning of a second piece of text. This simple abstraction of an exceedingly complex problem has broad appeal partly because it can be conceived also as a component in other NLP applications, from Machine Translation to Semantic Search to Information Extraction. It also avoids commitment to any specific meaning representation and reasoning framework, broadening its appeal within the research community. This level of abstraction also facilitates evaluation, a crucial component of any technological advancement program. This book explains the RTE task formulation adopted by the NLP research community, and gives a clear overview of research in this area. It draws out commonalities in this research, detailing the intuitions behind dominant approaches and their theoretical underpinnings. This book has been written with a wide audience in mind, but is intended to inform all readers about the state of the art in this fascinating field, to give a clear understanding of the principles underlying RTE research to date, and to highlight the short- and long-term research goals that will advance this technology.
This book constitutes the joint refereed proceedings of the 20th Symposium on the Integration of Symbolic Computation and Mechanized Reasoning, Calculemus 2013, 6th International Workshop on Digital Mathematics Libraries, DML 2013, Systems and Projects, held in Bath, UK as part of CICM 2013, the Conferences on Intelligent Computer Mathematics. The 7 revised full papers out of 18 submissions for MKM 2013, 5 revised full papers out of 12 submissions for Calculemus 2013, 6 revised full papers out of 8 submissions for DML 2013, and 12 revised full papers out of 16 submissions for Systems and Project track presented together with 3 invited talks were carefully reviewed and selected, resulting in 33 papers from a total of 73 submissions.
Many NLP tasks have at their core a subtask of extracting the dependencies-who did what to whom-from natural language sentences. This task can be understood as the inverse of the problem solved in different ways by diverse human languages, namely, how to indicate the relationship between different parts of a sentence. Understanding how languages solve the problem can be extremely useful in both feature design and error analysis in the application of machine learning to NLP. Likewise, understanding cross-linguistic variation can be important for the design of MT systems and other multilingual applications. The purpose of this book is to present in a succinct and accessible fashion information about the morphological and syntactic structure of human languages that can be useful in creating more linguistically sophisticated, more language-independent, and thus more successful NLP systems. Table of Contents: Acknowledgments / Introduction/motivation / Morphology: Introduction / Morphophonology / Morphosyntax / Syntax: Introduction / Parts of speech / Heads, arguments, and adjuncts / Argument types and grammatical functions / Mismatches between syntactic position and semantic roles / Resources / Bibliography / Author's Biography / General Index / Index of Languages
This book constitutes the proceedings of the 6th International Conference on Nonlinear Speech Processing, NOLISP 2013, held in Mons, Belgium, in June 2013. The 27 refereed papers included in this volume were carefully reviewed and selected from 34 submissions. The paper are organized in topical sections on speech and audio analysis; speech synthesis; speech-based biomedical applications; automatic speech recognition; and speech enhancement.
This book introduces basic supervised learning algorithms applicable to natural language processing (NLP) and shows how the performance of these algorithms can often be improved by exploiting the marginal distribution of large amounts of unlabeled data. One reason for that is data sparsity, i.e., the limited amounts of data we have available in NLP. However, in most real-world NLP applications our labeled data is also heavily biased. This book introduces extensions of supervised learning algorithms to cope with data sparsity and different kinds of sampling bias. This book is intended to be both readable by first-year students and interesting to the expert audience. My intention was to introduce what is necessary to appreciate the major challenges we face in contemporary NLP related to data sparsity and sampling bias, without wasting too much time on details about supervised learning algorithms or particular NLP applications. I use text classification, part-of-speech tagging, and dependency parsing as running examples, and limit myself to a small set of cardinal learning algorithms. I have worried less about theoretical guarantees ("this algorithm never does too badly") than about useful rules of thumb ("in this case this algorithm may perform really well"). In NLP, data is so noisy, biased, and non-stationary that few theoretical guarantees can be established and we are typically left with our gut feelings and a catalogue of crazy ideas. I hope this book will provide its readers with both. Throughout the book we include snippets of Python code and empirical evaluations, when relevant. |
You may like...
Foundation Models for Natural Language…
Gerhard PaaĆ, Sven Giesselbach
Hardcover
R884
Discovery Miles 8 840
Annotation, Exploitation and Evaluation…
Silvia Hansen-Schirra, Sambor Grucza
Hardcover
R919
Discovery Miles 9 190
|