![]() |
![]() |
Your cart is empty |
||
Books > Computing & IT > Applications of computing > Artificial intelligence > Natural language & machine translation
This book investigates the nature of generalization in language and
examines how language is known by adults and acquired by children.
It looks at how and why constructions are learned, the relation
between their forms and functions, and how cross-linguistic and
language-internal
This open access book provides a comprehensive overview of the state of the art in research and applications of Foundation Models and is intended for readers familiar with basic Natural Language Processing (NLP) concepts. Over the recent years, a revolutionary new paradigm has been developed for training models for NLP. These models are first pre-trained on large collections of text documents to acquire general syntactic knowledge and semantic information. Then, they are fine-tuned for specific tasks, which they can often solve with superhuman accuracy. When the models are large enough, they can be instructed by prompts to solve new tasks without any fine-tuning. Moreover, they can be applied to a wide range of different media and problem domains, ranging from image and video processing to robot control learning. Because they provide a blueprint for solving many tasks in artificial intelligence, they have been called Foundation Models. After a brief introduction to basic NLP models the main pre-trained language models BERT, GPT and sequence-to-sequence transformer are described, as well as the concepts of self-attention and context-sensitive embedding. Then, different approaches to improving these models are discussed, such as expanding the pre-training criteria, increasing the length of input texts, or including extra knowledge. An overview of the best-performing models for about twenty application areas is then presented, e.g., question answering, translation, story generation, dialog systems, generating images from text, etc. For each application area, the strengths and weaknesses of current models are discussed, and an outlook on further developments is given. In addition, links are provided to freely available program code. A concluding chapter summarizes the economic opportunities, mitigation of risks, and potential developments of AI.
The communication field is evolving rapidly in order to keep up with society's demands. As such, it becomes imperative to research and report recent advancements in computational intelligence as it applies to communication networks. The Handbook of Research on Recent Developments in Intelligent Communication Application is a pivotal reference source for the latest developments on emerging data communication applications. Featuring extensive coverage across a range of relevant perspectives and topics, such as satellite communication, cognitive radio networks, and wireless sensor networks, this book is ideally designed for engineers, professionals, practitioners, upper-level students, and academics seeking current information on emerging communication networking trends.
Information in today's advancing world is rapidly expanding and becoming widely available. This eruption of data has made handling it a daunting and time-consuming task. Natural language processing (NLP) is a method that applies linguistics and algorithms to large amounts of this data to make it more valuable. NLP improves the interaction between humans and computers, yet there remains a lack of research that focuses on the practical implementations of this trending approach. Neural Networks for Natural Language Processing is a collection of innovative research on the methods and applications of linguistic information processing and its computational properties. This publication will support readers with performing sentence classification and language generation using neural networks, apply deep learning models to solve machine translation and conversation problems, and apply deep structured semantic models on information retrieval and natural language applications. While highlighting topics including deep learning, query entity recognition, and information retrieval, this book is ideally designed for research and development professionals, IT specialists, industrialists, technology developers, data analysts, data scientists, academics, researchers, and students seeking current research on the fundamental concepts and techniques of natural language processing.
Language-that is, oral or written content that references abstract concepts in subtle ways-is what sets us apart as a species, and in an age defined by such content, language has become both the fuel and the currency of our modern information society. This has posed a vexing new challenge for linguists and engineers working in the field of language-processing: how do we parse and process not just language itself, but language in vast, overwhelming quantities? Modern Computational Models of Semantic Discovery in Natural Language compiles and reviews the most prominent linguistic theories into a single source that serves as an essential reference for future solutions to one of the most important challenges of our age. This comprehensive publication benefits an audience of students and professionals, researchers, and practitioners of linguistics and language discovery. This book includes a comprehensive range of topics and chapters covering digital media, social interaction in online environments, text and data mining, language processing and translation, and contextual documentation, among others.
This book covers theoretical work, applications, approaches, and techniques for computational models of information and its presentation by language (artificial, human, or natural in other ways). Computational and technological developments that incorporate natural language are proliferating. Adequate coverage encounters difficult problems related to ambiguities and dependency on context and agents (humans or computational systems). The goal is to promote computational systems of intelligent natural language processing and related models of computation, language, thought, mental states, reasoning, and other cognitive processes.
These proceedings presents the state-of-the-art in spoken dialog systems with applications in robotics, knowledge access and communication. It addresses specifically: 1. Dialog for interacting with smartphones; 2. Dialog for Open Domain knowledge access; 3. Dialog for robot interaction; 4. Mediated dialog (including crosslingual dialog involving Speech Translation); and,5. Dialog quality evaluation. These articles were presented at the IWSDS 2012 workshop.
When viewed through a political lens, the act of defining terms in natural language arguably transforms knowledge into values. This unique volume explores how corporate, military, academic, and professional values shaped efforts to define computer terminology and establish an information engineering profession as a precursor to what would become computer science. As the Cold War heated up, U.S. federal agencies increasingly funded university researchers and labs to develop technologies, like the computer, that would ensure that the U.S. maintained economic prosperity and military dominance over the Soviet Union. At the same time, private corporations saw opportunities for partnering with university labs and military agencies to generate profits as they strengthened their business positions in civilian sectors. They needed a common vocabulary and principles of streamlined communication to underpin the technology development that would ensure national prosperity and military dominance. investigates how language standardization contributed to the professionalization of computer science as separate from mathematics, electrical engineering, and physics examines traditions of language standardization in earlier eras of rapid technology development around electricity and radio highlights the importance of the analogy of "the computer is like a human" to early explanations of computer design and logic traces design and development of electronic computers within political and economic contexts foregrounds the importance of human relationships in decisions about computer design This in-depth humanistic study argues for the importance of natural language in shaping what people come to think of as possible and impossible relationships between computers and humans. The work is a key reference in the history of technology and serves as a source textbook on the human-level history of computing. In addition, it addresses those with interests in sociolinguistic questions around technology studies, as well as technology development at the nexus of politics, business, and human relations.
This book provides a corpus-led analysis of multi-word units (MWUs) in English, specifically fixed pairs of nouns which are linked by a conjunction, such as 'mum and dad', 'bride and groom' and 'law and order'. Crucially, the occurrence pattern of such pairs is dependent on genre, and this book aims to document the structural distribution of some key Linked Noun Groups (LNGs). The author looks at the usage patterns found in a range of poetry and fiction dating from the 17th to 20th century, and also highlights the important role such binomials play in academic English, while acknowledging that they are far less common in casual spoken English. His findings will be highly relevant to students and scholars working in language teaching, stylistics, and language technology (including AI).
Build and deploy intelligent applications for natural language processing with Python by using industry standard tools and recently popular methods in deep learning Key Features A no-math, code-driven programmer's guide to text processing and NLP Get state of the art results with modern tooling across linguistics, text vectors and machine learning Fundamentals of NLP methods from spaCy, gensim, scikit-learn and PyTorch Book DescriptionNLP in Python is among the most sought after skills among data scientists. With code and relevant case studies, this book will show how you can use industry-grade tools to implement NLP programs capable of learning from relevant data. We will explore many modern methods ranging from spaCy to word vectors that have reinvented NLP. The book takes you from the basics of NLP to building text processing applications. We start with an introduction to the basic vocabulary along with a workflow for building NLP applications. We use industry-grade NLP tools for cleaning and pre-processing text, automatic question and answer generation using linguistics, text embedding, text classifier, and building a chatbot. With each project, you will learn a new concept of NLP. You will learn about entity recognition, part of speech tagging and dependency parsing for Q and A. We use text embedding for both clustering documents and making chatbots, and then build classifiers using scikit-learn. We conclude by deploying these models as REST APIs with Flask. By the end, you will be confident building NLP applications, and know exactly what to look for when approaching new challenges. What you will learn Understand classical linguistics in using English grammar for automatically generating questions and answers from a free text corpus Work with text embedding models for dense number representations of words, subwords and characters in the English language for exploring document clustering Deep Learning in NLP using PyTorch with a code-driven introduction to PyTorch Using an NLP project management Framework for estimating timelines and organizing your project into stages Hack and build a simple chatbot application in 30 minutes Deploy an NLP or machine learning application using Flask as RESTFUL APIs Who this book is forProgrammers who wish to build systems that can interpret language. Exposure to Python programming is required. Familiarity with NLP or machine learning vocabulary will be helpful, but not mandatory.
The concept of natural language processing has become one of the preferred methods to better understand consumers, especially in recent years when digital technologies and research methods have developed exponentially. It has become apparent that when responding to international consumers through multiple platforms and speaking in the same language in which the consumers express themselves, companies are improving their standings within the public sphere. Natural Language Processing for Global and Local Business provides research exploring the theoretical and practical phenomenon of natural language processing through different languages and platforms in terms of today's conditions. Featuring coverage on a broad range of topics such as computational linguistics, information engineering, and translation technology, this book is ideally designed for IT specialists, academics, researchers, students, and business professionals seeking current research on improving and understanding the consumer experience.
This new textbook examines the motivations and the different algorithms for automatic document summarization (ADS). We performed a recent state of the art. The book shows the main problems of ADS, difficulties and the solutions provided by the community. It presents recent advances in ADS, as well as current applications and trends. The approaches are statistical, linguistic and symbolic. Several exemples are included in order to clarify the theoretical concepts. The books currently available in the area of Automatic Document Summarization are not recent. Powerful algorithms have been developed in recent years that include several applications of ADS. The development of recent technology has impacted on the development of algorithms and their applications. The massive use of social networks and the new forms of the technology requires the adaptation of the classical methods of text summarizers. This is a new textbook on Automatic Text Summarization, based on teaching materials used in two or one-semester courses. It presents a extensive state-of-art and describes the new systems on the subject. Previous automatic summarization books have been either collections of specialized papers, or else authored books with only a chapter or two devoted to the field as a whole. In other hand, the classic books on the subject are not recent.
To sustain and stay at the top of the market and give absolute comfort to the consumers, industries are using different strategies and technologies. Natural language processing (NLP) is a technology widely penetrating the market, irrespective of the industry and domains. It is extensively applied in businesses today, and it is the buzzword in every engineer's life. NLP can be implemented in all those areas where artificial intelligence is applicable either by simplifying the communication process or by refining and analyzing information. Neural machine translation has improved the imitation of professional translations over the years. When applied in neural machine translation, NLP helps educate neural machine networks. This can be used by industries to translate low-impact content including emails, regulatory texts, etc. Such machine translation tools speed up communication with partners while enriching other business interactions. Deep Natural Language Processing and AI Applications for Industry 5.0 provides innovative research on the latest findings, ideas, and applications in fields of interest that fall under the scope of NLP including computational linguistics, deep NLP, web analysis, sentiments analysis for business, and industry perspective. This book covers a wide range of topics such as deep learning, deepfakes, text mining, blockchain technology, and more, making it a crucial text for anyone interested in NLP and artificial intelligence, including academicians, researchers, professionals, industry experts, business analysts, data scientists, data analysts, healthcare system designers, intelligent system designers, practitioners, and students.
Describing the technologies to combine language resources flexibly as web services, this book provides valuable case studies for those who work in services computing, language resources, human-computer interaction (HCI), computer-supported cooperative work (CSCW), and service science. The authors have been operating the Language Grid, which wraps existing language resources as atomic language services and enables users to compose new services by combining them. From architecture level to service composition level, the book explains how to resolve infrastructural and operational difficulties in sharing and combining language resources, including interoperability of language service infrastructures, various types of language service policies, human services, and service failures.The research based on the authors' operating experiences of handling complicated issues such as intellectual property and interoperability of language resources contributes to exploitation of language resources as a service. On the other hand, both the analysis based on using services and the design of new services can bring significant results. A new style of multilingual communication supported by language services is worthy of analysis in HCI/CSCW, and the design process of language services is the focus of valuable case studies in service science. By using language resources in different ways based on the Language Grid, many activities are highly regarded by diverse communities. This book consists of four parts: (1) two types of language service platforms to interconnect language services across service grids, (2) various language service composition technologies that improve the reusability, efficiency, and accuracy of composite services, (3) research work and activities in creating language resources and services, and (4) various applications and tools for understanding and designing language services that well support intercultural collaboration.
This book presents methods and approaches used to identify the true author of a doubtful document or text excerpt. It provides a broad introduction to all text categorization problems (like authorship attribution, psychological traits of the author, detecting fake news, etc.) grounded in stylistic features. Specifically, machine learning models as valuable tools for verifying hypotheses or revealing significant patterns hidden in datasets are presented in detail. Stylometry is a multi-disciplinary field combining linguistics with both statistics and computer science. The content is divided into three parts. The first, which consists of the first three chapters, offers a general introduction to stylometry, its potential applications and limitations. Further, it introduces the ongoing example used to illustrate the concepts discussed throughout the remainder of the book. The four chapters of the second part are more devoted to computer science with a focus on machine learning models. Their main aim is to explain machine learning models for solving stylometric problems. Several general strategies used to identify, extract, select, and represent stylistic markers are explained. As deep learning represents an active field of research, information on neural network models and word embeddings applied to stylometry is provided, as well as a general introduction to the deep learning approach to solving stylometric questions. In turn, the third part illustrates the application of the previously discussed approaches in real cases: an authorship attribution problem, seeking to discover the secret hand behind the nom de plume Elena Ferrante, an Italian writer known worldwide for her My Brilliant Friend's saga; author profiling in order to identify whether a set of tweets were generated by a bot or a human being and in this second case, whether it is a man or a woman; and an exploration of stylistic variations over time using US political speeches covering a period of ca. 230 years. A solutions-based approach is adopted throughout the book, and explanations are supported by examples written in R. To complement the main content and discussions on stylometric models and techniques, examples and datasets are freely available at the author's Github website.
Natural language processing (NLP) is a branch of artificial intelligence that has emerged as a prevalent method of practice for a sizeable amount of companies. NLP enables software to understand human language and process complex data that is generated within businesses. In a competitive market, leading organizations are showing an increased interest in implementing this technology to improve user experience and establish smarter decision-making methods. Research on the application of intelligent analytics is crucial for professionals and companies who wish to gain an edge on the opposition. The Handbook of Research on Natural Language Processing and Smart Service Systems is a collection of innovative research on the integration and development of intelligent software tools and their various applications within professional environments. While highlighting topics including discourse analysis, information retrieval, and advanced dialog systems, this book is ideally designed for developers, practitioners, researchers, managers, engineers, academicians, business professionals, scholars, policymakers, and students seeking current research on the improvement of competitive practices through the use of NLP and smart service systems.
|
![]() ![]() You may like...
Principles of Intelligent Automobiles
Xiubin Zhang, Muhammad Mansoor Khan
Hardcover
R4,605
Discovery Miles 46 050
Adaptive and Intelligent Control of…
Ravi Patel, Dipankar Deb, …
Hardcover
R3,020
Discovery Miles 30 200
Smart Computing Techniques and…
Suresh Chandra Satapathy, Vikrant Bhateja, …
Hardcover
R6,082
Discovery Miles 60 820
Informatics in Control, Automation and…
Oleg Gusikhin, Kurosh Madani
Hardcover
R5,988
Discovery Miles 59 880
Fuzzy Modeling and Control: Theory and…
Fernando Matia, G. Nicolas Marichal, …
Hardcover
R5,089
Discovery Miles 50 890
Adex Optimized Adaptive Controllers and…
Juan M. Martin-Sanchez, Jose Rodellar
Hardcover
R4,227
Discovery Miles 42 270
Electro-Mechanical Actuators for the…
Mirko Mazzoleni, Gianpietro Di Rito, …
Hardcover
R5,025
Discovery Miles 50 250
|