![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Artificial intelligence > Natural language & machine translation
It has been traditional in phonetic research to characterize monophthongs using a set of static formant frequencies, i.e., formant frequencies taken from a single time-point in the vowel or averaged over the time-course of the vowel. However, over the last twenty years a growing body of research has demonstrated that, at least for a number of dialects of North American English, vowels which are traditionally described as monophthongs often have substantial spectral change. Vowel inherent spectral change has been observed in speakers' productions, and has also been found to have a substantial effect on listeners' perception. In terms of acoustics, the traditional categorical distinction between monophthongs and diphthongs can be replaced by a gradient description of dynamic spectral patterns. This book includes chapters addressing various aspects of vowel inherent spectral change (VISC), including theoretical and experimental studies of the perceptually relevant aspects of VISC, the relationship between articulation (vocal-tract trajectories) and VISC, historical changes related VISC, cross-dialect, cross-language, and cross-age-group comparisons of VISC, the effects of VISC on second-language speech learning, and the use of VISC in forensic voice comparison.
The explosion of information technology has led to substantial growth of web-accessible linguistic data in terms of quantity, diversity and complexity. These resources become even more useful when interlinked with each other to generate network effects. The general trend of providing data online is thus accompanied by newly developing methodologies to interconnect linguistic data and metadata. This includes linguistic data collections, general-purpose knowledge bases (e.g., the DBpedia, a machine-readable edition of the Wikipedia), and repositories with specific information about languages, linguistic categories and phenomena. The Linked Data paradigm provides a framework for interoperability and access management, and thereby allows to integrate information from such a diverse set of resources. The contributions assembled in this volume illustrate the band-width of applications of the Linked Data paradigm for representative types of language resources. They cover lexical-semantic resources, annotated corpora, typological databases as well as terminology and metadata repositories. The book includes representative applications from diverse fields, ranging from academic linguistics (e.g., typology and corpus linguistics) over applied linguistics (e.g., lexicography and translation studies) to technical applications (in computational linguistics, Natural Language Processing and information technology). This volume accompanies the Workshop on Linked Data in Linguistics 2012 (LDL-2012) in Frankfurt/M., Germany, organized by the Open Linguistics Working Group (OWLG) of the Open Knowledge Foundation (OKFN). It assembles contributions of the workshop participants and, beyond this, it summarizes initial steps in the formation of a Linked Open Data cloud of linguistic resources, the Linguistic Linked Open Data cloud (LLOD).
Dialect Accent Features for Establishing Speaker Identity: A Case Study discusses the subject of forensic voice identification and speaker profiling. Specifically focusing on speaker profiling and using dialects of the Hindi language, widely used in India, the authors have contributed to the body of research on speaker identification by using accent feature as the discriminating factor. This case study contributes to the understanding of the speaker identification process in a situation where unknown speech samples are in different language/dialect than the recording of a suspect. The authors' data establishes that vowel quality, quantity, intonation and tone of a speaker as compared to Khariboli (standard Hindi) could be the potential features for identification of dialect accent.
It is becoming crucial to accurately estimate and monitor speech quality in various ambient environments to guarantee high quality speech communication. This practical hands-on book shows speech intelligibility measurement methods so that the readers can start measuring or estimating speech intelligibility of their own system. The book also introduces subjective and objective speech quality measures, and describes in detail speech intelligibility measurement methods. It introduces a diagnostic rhyme test which uses rhyming word-pairs, and includes: An investigation into the effect of word familiarity on speech intelligibility. Speech intelligibility measurement of localized speech in virtual 3-D acoustic space using the rhyme test. Estimation of speech intelligibility using objective measures, including the ITU standard PESQ measures, and automatic speech recognizers.
Novel Techniques for Dialectal Arabic Speech describes approaches to improve automatic speech recognition for dialectal Arabic. Since speech resources for dialectal Arabic speech recognition are very sparse, the authors describe how existing Modern Standard Arabic (MSA) speech data can be applied to dialectal Arabic speech recognition, while assuming that MSA is always a second language for all Arabic speakers. In this book, Egyptian Colloquial Arabic (ECA) has been chosen as a typical Arabic dialect. ECA is the first ranked Arabic dialect in terms of number of speakers, and a high quality ECA speech corpus with accurate phonetic transcription has been collected. MSA acoustic models were trained using news broadcast speech. In order to cross-lingually use MSA in dialectal Arabic speech recognition, the authors have normalized the phoneme sets for MSA and ECA. After this normalization, they have applied state-of-the-art acoustic model adaptation techniques like Maximum Likelihood Linear Regression (MLLR) and Maximum A-Posteriori (MAP) to adapt existing phonemic MSA acoustic models with a small amount of dialectal ECA speech data. Speech recognition results indicate a significant increase in recognition accuracy compared to a baseline model trained with only ECA data.
The Social Web (including services such as MySpace, Flickr, last.fm, and WordPress) has captured the attention of millions of users as well as billions of dollars in investment and acquisition. Social websites, evolving around the connections between people and their objects of interest, are encountering boundaries in the areas of information integration, dissemination, reuse, portability, searchability, automation and demanding tasks like querying. The Semantic Web is an ideal platform for interlinking and performing operations on diverse person- and object-related data available from the Social Web, and has produced a variety of approaches to overcome the boundaries being experienced in Social Web application areas. After a short overview of both the Social Web and the Semantic Web, Breslin et al. describe some popular social media and social networking applications, list their strengths and limitations, and describe some applications of Semantic Web technology to address their current shortcomings by enhancing them with semantics. Across these social websites, they demonstrate a twofold approach for interconnecting the islands that are social websites with semantic technologies, and for powering semantic applications with rich community-created content. They conclude with observations on how the application of Semantic Web technologies to the Social Web is leading towards the "Social Semantic Web" (sometimes also called "Web 3.0"), forming a network of interlinked and semantically-rich content and knowledge. The book is intended for computer science professionals, researchers, and graduates interested in understanding the technologies and research issues involved in applying Semantic Web technologies to social software. Practitioners and developers interested in applications such as blogs, social networks or wikis will also learn about methods for increasing the levels of automation in these forms of Web communication.
This volume includes proceedings articles presented at the Workshop on Paralinguistic Information and its Integration in Spoken Dialogue Systems held in Granada, Spain. The material focuses on the three broad areas of spoken dialogue systems for robotics, emotions and spoken dialogue systems, and Spoken dialogue systems for real-world applications The workshop proceedings are part of the 3rd Annual International Workshop on Spoken Dialogue Systems, which brings together researchers from all over the world working in the field of spoken dialogue systems. It provides an international forum for the presentation of research and applications, and for lively discussions among researchers as well as industrialists.
This book provides extensive insight into the possibilities and challenges of XML in building new information management solutions in networked organizations. After a brief introduction to Web communication features and XML fundamentals, the book examines the benefits of adopting XML and illustrates various types of XML use: XML in document management; XML for data-centric and multimedia components; XML as a format for metadata, including metadata for the Semantic Web; and XML in support of data interchange between software applications and among organizations. The challenges of adopting XML in large-scale information management are also discussed. In addition, applications across a broad spectrum are examined and numerous case studies pertaining to the adoption of XML are presented. The book is particularly suitable for courses offered in Information Studies, Information Systems, or Information Technology. It also serves as an excellent practical guide for professionals in information management and provides important support material for courses in Computer Science and in Business.
The proceedings includes cutting-edge research articles from the Fourth International Conference on Signal and Image Processing (ICSIP), which is organised by Dr. N.G.P. Institute of Technology, Kalapatti, Coimbatore. The Conference provides academia and industry to discuss and present the latest technological advances and research results in the fields of theoretical, experimental, and application of signal, image and video processing. The book provides latest and most informative content from engineers and scientists in signal, image and video processing from around the world, which will benefit the future research community to work in a more cohesive and collaborative way.
The proceedings includes cutting-edge research articles from the Fourth International Conference on Signal and Image Processing (ICSIP), which is organised by Dr. N.G.P. Institute of Technology, Kalapatti, Coimbatore. The Conference provides academia and industry to discuss and present the latest technological advances and research results in the fields of theoretical, experimental, and application of signal, image and video processing. The book provides latest and most informative content from engineers and scientists in signal, image and video processing from around the world, which will benefit the future research community to work in a more cohesive and collaborative way.
The past decade has seen a revolution in the field of spoken dialogue systems. As in other areas of Computer Science and Artificial Intelligence, data-driven methods are now being used to drive new methodologies for system development and evaluation. This book is a unique contribution to that ongoing change. A new methodology for developing spoken dialogue systems is described in detail. The journey starts and ends with human behaviour in interaction, and explores methods for learning from the data, for building simulation environments for training and testing systems, and for evaluating the results. The detailed material covers: Spoken and Multimodal dialogue systems, Wizard-of-Oz data collection, User Simulation methods, Reinforcement Learning, and Evaluation methodologies. The book is a research guide for students and researchers with a background in Computer Science, AI, or Machine Learning. It navigates through a detailed case study in data-driven methods for development and evaluation of spoken dialogue systems. Common challenges associated with this approach are discussed and example solutions are provided. This work provides insights, lessons, and inspiration for future research and development - not only for spoken dialogue systems in particular, but for data-driven approaches to human-machine interaction in general.
The impact of computer systems that can understand natural language will be tremendous. To develop this capability we need to be able to automatically and efficiently analyze large amounts of text. Manually devised rules are not sufficient to provide coverage to handle the complex structure of natural language, necessitating systems that can automatically learn from examples. To handle the flexibility of natural language, it has become standard practice to use statistical models, which assign probabilities for example to the different meanings of a word or the plausibility of grammatical constructions. This book develops a general coarse-to-fine framework for learning and inference in large statistical models for natural language processing. Coarse-to-fine approaches exploit a sequence of models which introduce complexity gradually. At the top of the sequence is a trivial model in which learning and inference are both cheap. Each subsequent model refines the previous one, until a final, full-complexity model is reached. Applications of this framework to syntactic parsing, speech recognition and machine translation are presented, demonstrating the effectiveness of the approach in terms of accuracy and speed. The book is intended for students and researchers interested in statistical approaches to Natural Language Processing. "Slav s work"Coarse-to-Fine Natural Language Processing "represents a major advance in the area of syntactic parsing, and a great advertisement for the superiority of the machine-learning approach." Eugene Charniak (Brown University)"
Two Top Industry Leaders Speak Out Judith Markowitz When Amy asked me to co-author the foreword to her new book on advances in speech recognition, I was honored. Amy's work has always been infused with c- ative intensity, so I knew the book would be as interesting for established speech professionals as for readers new to the speech-processing industry. The fact that I would be writing the foreward with Bill Scholz made the job even more enjoyable. Bill and I have known each other since he was at UNISYS directing projects that had a profound impact on speech-recognition tools and applications. Bill Scholz The opportunity to prepare this foreword with Judith provides me with a rare oppor- nity to collaborate with a seasoned speech professional to identify numerous signi- cant contributions to the field offered by the contributors whom Amy has recruited. Judith and I have had our eyes opened by the ideas and analyses offered by this collection of authors. Speech recognition no longer needs be relegated to the ca- gory of an experimental future technology; it is here today with sufficient capability to address the most challenging of tasks. And the point-click-type approach to GUI control is no longer sufficient, especially in the context of limitations of mode- day hand held devices. Instead, VUI and GUI are being integrated into unified multimodal solutions that are maturing into the fundamental paradigm for comput- human interaction in the future.
This book constitutes the thoroughly refereed proceedings of the 9th Italian Research Conference on Digital Libraries, held in Rome, Italy, in January/February 2013. The 18 full papers presented together with an invited paper and a panel paper were selected from extended versions of the presentations given at the conference. The papers then went through an additional round of reviewing and revision after the event. The papers are organized in topical sections on information access; Digital Library (DL) architecture; DL projects; semantics and DLs; models and evaluation for DLs; DL applications; discussing DL perspectives.
This proceedings is a representation of decades of reasearch, teaching and application in the field. Image Processing, Fusion and Information Technology areas, Digital radio Communication, Wimax, Electrical engg, VLSI approach to processor design, embedded systems design are dealt in detail through models and illustrative techniques.
This book discusses the impact of spectral features extracted from frame level, glottal closure regions, and pitch-synchronous analysis on the performance of language identification systems. In addition to spectral features, the authors explore prosodic features such as intonation, rhythm, and stress features for discriminating the languages. They present how the proposed spectral and prosodic features capture the language specific information from two complementary aspects, showing how the development of language identification (LID) system using the combination of spectral and prosodic features will enhance the accuracy of identification as well as improve the robustness of the system. This book provides the methods to extract the spectral and prosodic features at various levels, and also suggests the appropriate models for developing robust LID systems according to specific spectral and prosodic features. Finally, the book discuss about various combinations of spectral and prosodic features, and the desired models to enhance the performance of LID systems.
Researchers in many disciplines have been concerned with modeling textual data in order to account for texts as the primary information unit of written communication. The book "Modelling, Learning and Processing of Text-Technological Data Structures" deals with this challenging information unit. It focuses on theoretical foundations of representing natural language texts as well as on concrete operations of automatic text processing. Following this integrated approach, the present volume includes contributions to a wide range of topics in the context of processing of textual data. This relates to the learning of ontologies from natural language texts, the annotation and automatic parsing of texts as well as the detection and tracking of topics in texts and hypertexts. In this way, the book brings together a wide range of approaches to procedural aspects of text technology as an emerging scientific discipline.
There is increasing interaction among communities with multiple languages, thus we need services that can effectively support multilingual communication. The Language Grid is an initiative to build an infrastructure that allows end users to create composite language services for intercultural collaboration. The aim is to support communities to create customized multilingual environments by using language services to overcome local language barriers. The stakeholders of the Language Grid are the language resource providers, the language service users, and the language grid operators who coordinate the former. This book includes 18 chapters in six parts that summarize various research results and associated development activities on the Language Grid. The chapters in Part I describe the framework of the Language Grid, i.e., service-oriented collective intelligence, used to bridge providers, users and operators. Two kinds of software are introduced, the service grid server software and the Language Grid Toolbox, and code for both is available via open source licenses. Part II describes technologies for service workflows that compose atomic language services. Part III reports on research work and activities relating to sharing and using language services. Part IV describes various applications of language services as applicable to intercultural collaboration. Part V contains reports on applying the Language Grid for translation activities, including localization of industrial documents and Wikipedia articles. Finally, Part VI illustrates how the Language Grid can be connected to other service grids, such as DFKI's Heart of Gold and smart classroom services in Tsinghua University in Beijing. The book will be valuable for researchers in artificial intelligence, natural language processing, services computing and human--computer interaction, particularly those who are interested in bridging technologies and user communities. "
Reasoning for Information: Seeking and Planning Dialogues provides a logic-based reasoning component for spoken language dialogue systems. This component, called Problem Assistant is responsible for processing constraints on a possible solution obtained from various sources, namely user and the system's domain-specific information. The authors also present findings on the implementation of a dialogue management interface to the Problem Assistant. The dialogue system supports simple mixed-initiative planning interactions in the TRAINS domain, which is still a relatively complex domain involving a number of logical constraints and relations forming the basis for the collaborative problem-solving behavior that drives the dialogue.
This book collects together several of the tutorials held at EUROGRAPHICS'89 in Hamburg. The conference was held under the motto "Integration, Visualisation, Interaction" and the tutorials reflect the conference theme. The Springer series EurographicSeminars with the volumes "Advances in Computer Graphics" regularly provides a professional update on current mainstream topics in the field. These publications give readers the opportunity to inform themselves thoroughly on the topics covered. The success of the series is mainly based on the expertise of the contributing authors, who are recognized professionals in their field. Starting out with one of the conference's main topics, the chapter "Visualization of Scientific Data" gives an overview of methods for displaying scientific results in an easily surveyable and comprehensible form. It presents algorithms and methods utilized to achieve visualization results in a form adequate for humans. User interfaces for such systems are also explored, and practical conclusions are drawn. The chapter "Color in Computer Graphics" describes the problems of manipulating and matching color in the real world. After some fundamental statements about color models and their relationships, the main emphasis is placed on the problem of objective color specification for computer graphics systems. It is very hard to match colors between devices such as scanners, printers and displays. Some suggestions on the effective use of color for graphics are also made.
DigiScript edits for production all digital manuscripts in the current industry formats PostScript and PDF, including pictures, text, graphics, color, and production parameters, independently of the hardware and software platforms used to prepare the document. Fundamental background information, technical know-how, and application examples from the professional prepress field are combined to help the user solve difficult production problems efficiently. The complete description of the DigiScript production environment allows the book to serve as a self-contained reference work. The accompanying CD-ROM provides all the data you need to test DigiScript on your NEXTSTEP 3.3 system as well as review copies of the examples introduced in the user handbook.
Originally entitled the "PostScript and Acrobat Bible" in German, this handbook achieves the seemingly impossible: it covers this tricky and technical field in an entertaining manner without getting bogged down in PostScript programming. It explains how several components work together and how to deal with real-world application and operating-system problems. The author genuinely wants to assist in overcoming cross-platform barriers using MS-DOS, Windows, Macintosh or UNIX and, accordingly, neither the book nor the tools are limited to one particular platform or operating system. The 9 chapters and 3 appendixes run the entire gamut, from the very basics right up to Ghostscript and the whole is creatively designed, making use of comical illustrations. In short, essential reading for all technically minded users of PostScript and Acrobat/PDF - from PC owners wanting to get the most out of their laser printers to graphic artists with Macs to system administrators and online publishers.
This book constitutes the refereed proceedings of the 12th China National Conference on Computational Linguistics, CCL 2013, and of the First International Symposium on Natural Language Processing Based on Naturally Annotated Big Data, NLP-NABD 2013, held in Suzhou, China, in October 2013. The 32 papers presented were carefully reviewed and selected from 252 submissions. The papers are organized in topical sections on word segmentation; open-domain question answering; discourse, coreference and pragmatics; statistical and machine learning methods in NLP; semantics; text mining, open-domain information extraction and machine reading of the Web; sentiment analysis, opinion mining and text classification; lexical semantics and ontologies; language resources and annotation; machine translation; speech recognition and synthesis; tagging and chunking; and large-scale knowledge acquisition and reasoning.
The digital age has had a profound effect on our cultural
heritage and the academic research that studies it. Staggering
amounts of objects, many of them of a textual nature, are being
digitised to make them more readily accessible to both experts and
laypersons. Besides a vast potential for more effective and
efficient preservation, management, and presentation, digitisation
offers opportunities to work with cultural heritage data in ways
that were never feasible or even imagined.
We are living in a multilingual world and the diversity in languages which are used to interact with information access systems has generated a wide variety of challenges to be addressed by computer and information scientists. The growing amount of non-English information accessible globally and the increased worldwide exposure of enterprises also necessitates the adaptation of Information Retrieval (IR) methods to new, multilingual settings. Peters, Braschler and Clough present a comprehensive description of the technologies involved in designing and developing systems for Multilingual Information Retrieval (MLIR). They provide readers with broad coverage of the various issues involved in creating systems to make accessible digitally stored materials regardless of the language(s) they are written in. Details on Cross-Language Information Retrieval (CLIR) are also covered that help readers to understand how to develop retrieval systems that cross language boundaries. Their work is divided into six chapters and accompanies the reader step-by-step through the various stages involved in building, using and evaluating MLIR systems. The book concludes with some examples of recent applications that utilise MLIR technologies. Some of the techniques described have recently started to appear in commercial search systems, while others have the potential to be part of future incarnations. The book is intended for graduate students, scholars, and practitioners with a basic understanding of classical text retrieval methods. It offers guidelines and information on all aspects that need to be taken into consideration when building MLIR systems, while avoiding too many 'hands-on details' that could rapidly become obsolete. Thus it bridges the gap between the material covered by most of the classical IR textbooks and the novel requirements related to the acquisition and dissemination of information in whatever language it is stored. |
You may like...
Intelligent Help Systems for UNIX
Stephen J. Hegner, Paul Mc Kevitt, …
Hardcover
R2,270
Discovery Miles 22 700
Linux - The Textbook, Second Edition
Syed Mansoor Sarwar, Robert M. Koretsky
Paperback
R1,351
Discovery Miles 13 510
|