|
Books > Computing & IT > Applications of computing > Artificial intelligence > Natural language & machine translation
This open access book provides a comprehensive overview of the
state of the art in research and applications of Foundation Models
and is intended for readers familiar with basic Natural Language
Processing (NLP) concepts. Over the recent years, a
revolutionary new paradigm has been developed for training models
for NLP. These models are first pre-trained on large collections of
text documents to acquire general syntactic knowledge and semantic
information. Then, they are fine-tuned for specific tasks, which
they can often solve with superhuman accuracy. When the models are
large enough, they can be instructed by prompts to solve new tasks
without any fine-tuning. Moreover, they can be applied to a wide
range of different media and problem domains, ranging from image
and video processing to robot control learning. Because they
provide a blueprint for solving many tasks in artificial
intelligence, they have been called Foundation Models. After
a brief introduction to basic NLP models the main pre-trained
language models BERT, GPT and sequence-to-sequence transformer are
described, as well as the concepts of self-attention and
context-sensitive embedding. Then, different approaches to
improving these models are discussed, such as expanding the
pre-training criteria, increasing the length of input texts, or
including extra knowledge. An overview of the best-performing
models for about twenty application areas is then presented, e.g.,
question answering, translation, story generation, dialog systems,
generating images from text, etc. For each application area, the
strengths and weaknesses of current models are discussed, and an
outlook on further developments is given. In addition, links are
provided to freely available program code. A concluding chapter
summarizes the economic opportunities, mitigation of risks, and
potential developments of AI.
The communication field is evolving rapidly in order to keep up
with society's demands. As such, it becomes imperative to research
and report recent advancements in computational intelligence as it
applies to communication networks. The Handbook of Research on
Recent Developments in Intelligent Communication Application is a
pivotal reference source for the latest developments on emerging
data communication applications. Featuring extensive coverage
across a range of relevant perspectives and topics, such as
satellite communication, cognitive radio networks, and wireless
sensor networks, this book is ideally designed for engineers,
professionals, practitioners, upper-level students, and academics
seeking current information on emerging communication networking
trends.
Language-that is, oral or written content that references abstract
concepts in subtle ways-is what sets us apart as a species, and in
an age defined by such content, language has become both the fuel
and the currency of our modern information society. This has posed
a vexing new challenge for linguists and engineers working in the
field of language-processing: how do we parse and process not just
language itself, but language in vast, overwhelming quantities?
Modern Computational Models of Semantic Discovery in Natural
Language compiles and reviews the most prominent linguistic
theories into a single source that serves as an essential reference
for future solutions to one of the most important challenges of our
age. This comprehensive publication benefits an audience of
students and professionals, researchers, and practitioners of
linguistics and language discovery. This book includes a
comprehensive range of topics and chapters covering digital media,
social interaction in online environments, text and data mining,
language processing and translation, and contextual documentation,
among others.
Information in today's advancing world is rapidly expanding and
becoming widely available. This eruption of data has made handling
it a daunting and time-consuming task. Natural language processing
(NLP) is a method that applies linguistics and algorithms to large
amounts of this data to make it more valuable. NLP improves the
interaction between humans and computers, yet there remains a lack
of research that focuses on the practical implementations of this
trending approach. Neural Networks for Natural Language Processing
is a collection of innovative research on the methods and
applications of linguistic information processing and its
computational properties. This publication will support readers
with performing sentence classification and language generation
using neural networks, apply deep learning models to solve machine
translation and conversation problems, and apply deep structured
semantic models on information retrieval and natural language
applications. While highlighting topics including deep learning,
query entity recognition, and information retrieval, this book is
ideally designed for research and development professionals, IT
specialists, industrialists, technology developers, data analysts,
data scientists, academics, researchers, and students seeking
current research on the fundamental concepts and techniques of
natural language processing.
This book provides a new multi-method, process-oriented approach
towards speech quality assessment, which allows readers to examine
the influence of speech transmission quality on a variety of
perceptual and cognitive processes in human listeners. Fundamental
concepts and methodologies surrounding the topic of
process-oriented quality assessment are introduced and discussed.
The book further describes a functional process model of human
quality perception, which theoretically integrates results obtained
in three experimental studies. This book's conceptual ideas,
empirical findings, and theoretical interpretations should be of
particular interest to researchers working in the fields of Quality
and Usability Engineering, Audio Engineering, Psychoacoustics,
Audiology, and Psychophysiology.
This book covers theoretical work, applications, approaches, and
techniques for computational models of information and its
presentation by language (artificial, human, or natural in other
ways). Computational and technological developments that
incorporate natural language are proliferating. Adequate coverage
encounters difficult problems related to ambiguities and dependency
on context and agents (humans or computational systems). The goal
is to promote computational systems of intelligent natural language
processing and related models of computation, language, thought,
mental states, reasoning, and other cognitive processes.
When viewed through a political lens, the act of defining terms in
natural language arguably transforms knowledge into values. This
unique volume explores how corporate, military, academic, and
professional values shaped efforts to define computer terminology
and establish an information engineering profession as a precursor
to what would become computer science. As the Cold War heated up,
U.S. federal agencies increasingly funded university researchers
and labs to develop technologies, like the computer, that would
ensure that the U.S. maintained economic prosperity and military
dominance over the Soviet Union. At the same time, private
corporations saw opportunities for partnering with university labs
and military agencies to generate profits as they strengthened
their business positions in civilian sectors. They needed a common
vocabulary and principles of streamlined communication to underpin
the technology development that would ensure national prosperity
and military dominance. investigates how language standardization
contributed to the professionalization of computer science as
separate from mathematics, electrical engineering, and physics
examines traditions of language standardization in earlier eras of
rapid technology development around electricity and radio
highlights the importance of the analogy of "the computer is like a
human" to early explanations of computer design and logic traces
design and development of electronic computers within political and
economic contexts foregrounds the importance of human relationships
in decisions about computer design This in-depth humanistic study
argues for the importance of natural language in shaping what
people come to think of as possible and impossible relationships
between computers and humans. The work is a key reference in the
history of technology and serves as a source textbook on the
human-level history of computing. In addition, it addresses those
with interests in sociolinguistic questions around technology
studies, as well as technology development at the nexus of
politics, business, and human relations.
This book provides a corpus-led analysis of multi-word units (MWUs)
in English, specifically fixed pairs of nouns which are linked by a
conjunction, such as 'mum and dad', 'bride and groom' and 'law and
order'. Crucially, the occurrence pattern of such pairs is
dependent on genre, and this book aims to document the structural
distribution of some key Linked Noun Groups (LNGs). The author
looks at the usage patterns found in a range of poetry and fiction
dating from the 17th to 20th century, and also highlights the
important role such binomials play in academic English, while
acknowledging that they are far less common in casual spoken
English. His findings will be highly relevant to students and
scholars working in language teaching, stylistics, and language
technology (including AI).
Build and deploy intelligent applications for natural language
processing with Python by using industry standard tools and
recently popular methods in deep learning Key Features A no-math,
code-driven programmer's guide to text processing and NLP Get state
of the art results with modern tooling across linguistics, text
vectors and machine learning Fundamentals of NLP methods from
spaCy, gensim, scikit-learn and PyTorch Book DescriptionNLP in
Python is among the most sought after skills among data scientists.
With code and relevant case studies, this book will show how you
can use industry-grade tools to implement NLP programs capable of
learning from relevant data. We will explore many modern methods
ranging from spaCy to word vectors that have reinvented NLP. The
book takes you from the basics of NLP to building text processing
applications. We start with an introduction to the basic vocabulary
along with a workflow for building NLP applications. We use
industry-grade NLP tools for cleaning and pre-processing text,
automatic question and answer generation using linguistics, text
embedding, text classifier, and building a chatbot. With each
project, you will learn a new concept of NLP. You will learn about
entity recognition, part of speech tagging and dependency parsing
for Q and A. We use text embedding for both clustering documents
and making chatbots, and then build classifiers using scikit-learn.
We conclude by deploying these models as REST APIs with Flask. By
the end, you will be confident building NLP applications, and know
exactly what to look for when approaching new challenges. What you
will learn Understand classical linguistics in using English
grammar for automatically generating questions and answers from a
free text corpus Work with text embedding models for dense number
representations of words, subwords and characters in the English
language for exploring document clustering Deep Learning in NLP
using PyTorch with a code-driven introduction to PyTorch Using an
NLP project management Framework for estimating timelines and
organizing your project into stages Hack and build a simple chatbot
application in 30 minutes Deploy an NLP or machine learning
application using Flask as RESTFUL APIs Who this book is
forProgrammers who wish to build systems that can interpret
language. Exposure to Python programming is required. Familiarity
with NLP or machine learning vocabulary will be helpful, but not
mandatory.
These proceedings presents the state-of-the-art in spoken dialog
systems with applications in robotics, knowledge access and
communication. It addresses specifically: 1. Dialog for interacting
with smartphones; 2. Dialog for Open Domain knowledge access; 3.
Dialog for robot interaction; 4. Mediated dialog (including
crosslingual dialog involving Speech Translation); and,5. Dialog
quality evaluation. These articles were presented at the IWSDS 2012
workshop.
The concept of natural language processing has become one of the
preferred methods to better understand consumers, especially in
recent years when digital technologies and research methods have
developed exponentially. It has become apparent that when
responding to international consumers through multiple platforms
and speaking in the same language in which the consumers express
themselves, companies are improving their standings within the
public sphere. Natural Language Processing for Global and Local
Business provides research exploring the theoretical and practical
phenomenon of natural language processing through different
languages and platforms in terms of today's conditions. Featuring
coverage on a broad range of topics such as computational
linguistics, information engineering, and translation technology,
this book is ideally designed for IT specialists, academics,
researchers, students, and business professionals seeking current
research on improving and understanding the consumer experience.
This book presents methods and approaches used to identify the true
author of a doubtful document or text excerpt. It provides a broad
introduction to all text categorization problems (like authorship
attribution, psychological traits of the author, detecting fake
news, etc.) grounded in stylistic features. Specifically, machine
learning models as valuable tools for verifying hypotheses or
revealing significant patterns hidden in datasets are presented in
detail. Stylometry is a multi-disciplinary field combining
linguistics with both statistics and computer science. The content
is divided into three parts. The first, which consists of the first
three chapters, offers a general introduction to stylometry, its
potential applications and limitations. Further, it introduces the
ongoing example used to illustrate the concepts discussed
throughout the remainder of the book. The four chapters of the
second part are more devoted to computer science with a focus on
machine learning models. Their main aim is to explain machine
learning models for solving stylometric problems. Several general
strategies used to identify, extract, select, and represent
stylistic markers are explained. As deep learning represents an
active field of research, information on neural network models and
word embeddings applied to stylometry is provided, as well as a
general introduction to the deep learning approach to solving
stylometric questions. In turn, the third part illustrates the
application of the previously discussed approaches in real cases:
an authorship attribution problem, seeking to discover the secret
hand behind the nom de plume Elena Ferrante, an Italian writer
known worldwide for her My Brilliant Friend's saga; author
profiling in order to identify whether a set of tweets were
generated by a bot or a human being and in this second case,
whether it is a man or a woman; and an exploration of stylistic
variations over time using US political speeches covering a period
of ca. 230 years. A solutions-based approach is adopted throughout
the book, and explanations are supported by examples written in R.
To complement the main content and discussions on stylometric
models and techniques, examples and datasets are freely available
at the author's Github website.
To sustain and stay at the top of the market and give absolute
comfort to the consumers, industries are using different strategies
and technologies. Natural language processing (NLP) is a technology
widely penetrating the market, irrespective of the industry and
domains. It is extensively applied in businesses today, and it is
the buzzword in every engineer's life. NLP can be implemented in
all those areas where artificial intelligence is applicable either
by simplifying the communication process or by refining and
analyzing information. Neural machine translation has improved the
imitation of professional translations over the years. When applied
in neural machine translation, NLP helps educate neural machine
networks. This can be used by industries to translate low-impact
content including emails, regulatory texts, etc. Such machine
translation tools speed up communication with partners while
enriching other business interactions. Deep Natural Language
Processing and AI Applications for Industry 5.0 provides innovative
research on the latest findings, ideas, and applications in fields
of interest that fall under the scope of NLP including
computational linguistics, deep NLP, web analysis, sentiments
analysis for business, and industry perspective. This book covers a
wide range of topics such as deep learning, deepfakes, text mining,
blockchain technology, and more, making it a crucial text for
anyone interested in NLP and artificial intelligence, including
academicians, researchers, professionals, industry experts,
business analysts, data scientists, data analysts, healthcare
system designers, intelligent system designers, practitioners, and
students.
Describing the technologies to combine language resources flexibly
as web services, this book provides valuable case studies for those
who work in services computing, language resources, human-computer
interaction (HCI), computer-supported cooperative work (CSCW), and
service science. The authors have been operating the Language Grid,
which wraps existing language resources as atomic language services
and enables users to compose new services by combining them. From
architecture level to service composition level, the book explains
how to resolve infrastructural and operational difficulties in
sharing and combining language resources, including
interoperability of language service infrastructures, various types
of language service policies, human services, and service
failures.The research based on the authors' operating experiences
of handling complicated issues such as intellectual property and
interoperability of language resources contributes to exploitation
of language resources as a service. On the other hand, both the
analysis based on using services and the design of new services can
bring significant results. A new style of multilingual
communication supported by language services is worthy of analysis
in HCI/CSCW, and the design process of language services is the
focus of valuable case studies in service science. By using
language resources in different ways based on the Language Grid,
many activities are highly regarded by diverse communities. This
book consists of four parts: (1) two types of language service
platforms to interconnect language services across service grids,
(2) various language service composition technologies that improve
the reusability, efficiency, and accuracy of composite services,
(3) research work and activities in creating language resources and
services, and (4) various applications and tools for understanding
and designing language services that well support intercultural
collaboration.
|
|