|
Books > Computing & IT > Applications of computing > Databases > Data capture & analysis
Quickly build and deploy machine learning models without managing
infrastructure, and improve productivity using Amazon SageMaker's
capabilities such as Amazon SageMaker Studio, Autopilot,
Experiments, Debugger, and Model Monitor Key Features Build, train,
and deploy machine learning models quickly using Amazon SageMaker
Analyze, detect, and receive alerts relating to various business
problems using machine learning algorithms and techniques Improve
productivity by training and fine-tuning machine learning models in
production Book DescriptionAmazon SageMaker enables you to quickly
build, train, and deploy machine learning (ML) models at scale,
without managing any infrastructure. It helps you focus on the ML
problem at hand and deploy high-quality models by removing the
heavy lifting typically involved in each step of the ML process.
This book is a comprehensive guide for data scientists and ML
developers who want to learn the ins and outs of Amazon SageMaker.
You'll understand how to use various modules of SageMaker as a
single toolset to solve the challenges faced in ML. As you
progress, you'll cover features such as AutoML, built-in algorithms
and frameworks, and the option for writing your own code and
algorithms to build ML models. Later, the book will show you how to
integrate Amazon SageMaker with popular deep learning libraries
such as TensorFlow and PyTorch to increase the capabilities of
existing models. You'll also learn to get the models to production
faster with minimum effort and at a lower cost. Finally, you'll
explore how to use Amazon SageMaker Debugger to analyze, detect,
and highlight problems to understand the current model state and
improve model accuracy. By the end of this Amazon book, you'll be
able to use Amazon SageMaker on the full spectrum of ML workflows,
from experimentation, training, and monitoring to scaling,
deployment, and automation. What you will learn Create and automate
end-to-end machine learning workflows on Amazon Web Services (AWS)
Become well-versed with data annotation and preparation techniques
Use AutoML features to build and train machine learning models with
AutoPilot Create models using built-in algorithms and frameworks
and your own code Train computer vision and NLP models using
real-world examples Cover training techniques for scaling, model
optimization, model debugging, and cost optimization Automate
deployment tasks in a variety of configurations using SDK and
several automation tools Who this book is forThis book is for
software engineers, machine learning developers, data scientists,
and AWS users who are new to using Amazon SageMaker and want to
build high-quality machine learning models without worrying about
infrastructure. Knowledge of AWS basics is required to grasp the
concepts covered in this book more effectively. Some understanding
of machine learning concepts and the Python programming language
will also be beneficial.
Get to grips with building and productionizing end-to-end big data
solutions in Azure and learn best practices for working with large
datasets Key Features Integrate with Azure Synapse Analytics,
Cosmos DB, and Azure HDInsight Kafka Cluster to scale and analyze
your projects and build pipelines Use Databricks SQL to run ad hoc
queries on your data lake and create dashboards Productionize a
solution using CI/CD for deploying notebooks and Azure Databricks
Service to various environments Book DescriptionAzure Databricks is
a unified collaborative platform for performing scalable analytics
in an interactive environment. The Azure Databricks Cookbook
provides recipes to get hands-on with the analytics process,
including ingesting data from various batch and streaming sources
and building a modern data warehouse. The book starts by teaching
you how to create an Azure Databricks instance within the Azure
portal, Azure CLI, and ARM templates. You'll work through clusters
in Databricks and explore recipes for ingesting data from sources,
including files, databases, and streaming sources such as Apache
Kafka and EventHub. The book will help you explore all the features
supported by Azure Databricks for building powerful end-to-end data
pipelines. You'll also find out how to build a modern data
warehouse by using Delta tables and Azure Synapse Analytics. Later,
you'll learn how to write ad hoc queries and extract meaningful
insights from the data lake by creating visualizations and
dashboards with Databricks SQL. Finally, you'll deploy and
productionize a data pipeline as well as deploy notebooks and Azure
Databricks service using continuous integration and continuous
delivery (CI/CD). By the end of this Azure book, you'll be able to
use Azure Databricks to streamline different processes involved in
building data-driven apps. What you will learn Read and write data
from and to various Azure resources and file formats Build a modern
data warehouse with Delta Tables and Azure Synapse Analytics
Explore jobs, stages, and tasks and see how Spark lazy evaluation
works Handle concurrent transactions and learn performance
optimization in Delta tables Learn Databricks SQL and create
real-time dashboards in Databricks SQL Integrate Azure DevOps for
version control, deploying, and productionizing solutions with
CI/CD pipelines Discover how to use RBAC and ACLs to restrict data
access Build end-to-end data processing pipeline for near real-time
data analytics Who this book is forThis recipe-based book is for
data scientists, data engineers, big data professionals, and
machine learning engineers who want to perform data analytics on
their applications. Prior experience of working with Apache Spark
and Azure is necessary to get the most out of this book.
Die Absicht, ein Buch iiber Programmieren von Ziffernrechenanlagen
zu schreiben, entstand auf Grund einer Vorlesung gleichen Titels,
die ich seit nunmehr sieben Jahren an der Technischen Hochschule
Wien halte. Ich hatte dabei bemerkt, daB das Interesse fiir die
Programmierung von Ziffernrechnern immer weitere Kreise zieht und
daB es moglich ist, dieses Interesse aus einem einheitlichen
Gesichtswinkel zu befriedigen. Der Zugang zur Kenntnis des
Programmierens erfolgt heute iiblicher- weise mit Hille der
Mathematischen Verfahrenstechnik oder von seiten der
Administrativen Automation, oder schlieBlich iiber die mit tech-
nischen Einzelheiten vermengte Beschreibung spezieller Maschinen.
Ich bin nun der Meinung, daB man ein Buch iiber Programmieren
schreiben kann, ohne auf Einzelheiten der Mathematischen
Verfahrenstechnik und der Biiroautomation oder auf technische
Eigenschaften spezieller Ma- schinen eingehen zu miissen, und ohne
damit jewells einem Tell der Leser das Verstandnis zu erschweren.
Was nach Fortlassung der ge- nannten Gebiete bleibt, ist nicht ein
trockener, unverstandlicher Rest, sondern der Inbegriff aller fiir
das Programmieren wesentlichen Prin- zipien. Sowohl der
Naturwissenschaftler als auch der Verwaltungsfach- mann, der diese
Prinzipien erfaBt hat, wird jederzeit in der Lage sein, sie seinen
besonderen Aufgaben dienstbar zu machen. Kapitel A solI zeigen,
welchen Platz der Rechenautomat unter den technischen
Errungenschaften einnimmt und wie er dorthin gelangt ist.
Besonderes Anliegen ist mir hier der geschichtliche Uberblick, well
einer- seits die deutschsprachigen Biicher auf diesem Gebiet kaum
historische Angaben enthalten und andererseits die
anglo-amerikanische Literatur die kontinentaleuropaische
Entwicklung iibergeht. - Kapitel B enthalt die Beschreibung einer
gedachten Maschine TElCO in allen Einzelheiten.
Think about your data intelligently and ask the right questions Key
Features Master data cleaning techniques necessary to perform
real-world data science and machine learning tasks Spot common
problems with dirty data and develop flexible solutions from first
principles Test and refine your newly acquired skills through
detailed exercises at the end of each chapter Book DescriptionData
cleaning is the all-important first step to successful data
science, data analysis, and machine learning. If you work with any
kind of data, this book is your go-to resource, arming you with the
insights and heuristics experienced data scientists had to learn
the hard way. In a light-hearted and engaging exploration of
different tools, techniques, and datasets real and fictitious,
Python veteran David Mertz teaches you the ins and outs of data
preparation and the essential questions you should be asking of
every piece of data you work with. Using a mixture of Python, R,
and common command-line tools, Cleaning Data for Effective Data
Science follows the data cleaning pipeline from start to end,
focusing on helping you understand the principles underlying each
step of the process. You'll look at data ingestion of a vast range
of tabular, hierarchical, and other data formats, impute missing
values, detect unreliable data and statistical anomalies, and
generate synthetic features. The long-form exercises at the end of
each chapter let you get hands-on with the skills you've acquired
along the way, also providing a valuable resource for academic
courses. What you will learn Ingest and work with common data
formats like JSON, CSV, SQL and NoSQL databases, PDF, and binary
serialized data structures Understand how and why we use tools such
as pandas, SciPy, scikit-learn, Tidyverse, and Bash Apply useful
rules and heuristics for assessing data quality and detecting bias,
like Benford's law and the 68-95-99.7 rule Identify and handle
unreliable data and outliers, examining z-score and other
statistical properties Impute sensible values into missing data and
use sampling to fix imbalances Use dimensionality reduction,
quantization, one-hot encoding, and other feature engineering
techniques to draw out patterns in your data Work carefully with
time series data, performing de-trending and interpolation Who this
book is forThis book is designed to benefit software developers,
data scientists, aspiring data scientists, teachers, and students
who work with data. If you want to improve your rigor in data
hygiene or are looking for a refresher, this book is for you. Basic
familiarity with statistics, general concepts in machine learning,
knowledge of a programming language (Python or R), and some
exposure to data science are helpful.
A comprehensive introduction to help you get up and running with
creating interactive dashboards to visualize and monitor
time-series data in no time Key Features Install, set up, and
configure Grafana for real-time data analysis and visualization
Visualize and monitor data using data sources such as InfluxDB,
Prometheus, and Elasticsearch Explore Grafana's multi-cloud support
with Microsoft Azure, Amazon CloudWatch, and Google Stackdriver
Book DescriptionGrafana is an open-source analytical platform used
to analyze and monitoring time-series data. This beginner's guide
will help you get to grips with Grafana's new features for
querying, visualizing, and exploring metrics and logs no matter
where they are stored. The book begins by showing you how to
install and set up the Grafana server. You'll explore the working
mechanism of various components of the Grafana interface along with
its security features, and learn how to visualize and monitor data
using, InfluxDB, Prometheus, Logstash, and Elasticsearch. This
Grafana book covers the advanced features of the Graph panel and
shows you how Stat, Table, Bar Gauge, and Text are used. You'll
build dynamic dashboards to perform end-to-end analytics and label
and organize dashboards into folders to make them easier to find.
As you progress, the book delves into the administrative aspects of
Grafana by creating alerts, setting permissions for teams, and
implementing user authentication. Along with exploring Grafana's
multi-cloud monitoring support, you'll also learn about Grafana
Loki, which is a backend logger for users running Prometheus and
Kubernetes. By the end of this book, you'll have gained all the
knowledge you need to start building interactive dashboards. What
you will learn Find out how to visualize data using Grafana
Understand how to work with the major components of the Graph panel
Explore mixed data sources, query inspector, and time interval
settings Discover advanced dashboard features such as annotations,
templating with variables, dashboard linking, and dashboard sharing
techniques Connect user authentication to Google, GitHub, and a
variety of external services Find out how Grafana can provide
monitoring support for cloud service infrastructures Who this book
is forThis book is for business intelligence developers, business
analysts, data analysts, and anyone interested in performing
time-series data analysis and monitoring using Grafana. Those
looking to create and share interactive dashboards or looking to
get up to speed with the latest features of Grafana will also find
this book useful. Although no prior knowledge of Grafana is
required, basic knowledge of data visualization and some experience
in Python programming will help you understand the concepts covered
in the book.
Build a continuously learning and adapting organization that can
extract increasing levels of business, customer and operational
value from the amalgamation of data and advanced analytics such as
AI and Machine Learning Key Features Master the Big Data Business
Model Maturity Index methodology to transition to a value-driven
organizational mindset Acquire implementable knowledge on digital
transformation through 8 practical laws Explore the economics
behind digital assets (data and analytics) that appreciate in value
when constructed and deployed correctly Book DescriptionIn today's
digital era, every organization has data, but just possessing
enormous amounts of data is not a sufficient market discriminator.
The Economics of Data, Analytics, and Digital Transformation aims
to provide actionable insights into the real market discriminators,
including an organization's data-fueled analytics products that
inspire innovation, deliver insights, help make practical
decisions, generate value, and produce mission success for the
enterprise. The book begins by first building your mindset to be
value-driven and introducing the Big Data Business Model Maturity
Index, its maturity index phases, and how to navigate the index.
You will explore value engineering, where you will learn how to
identify key business initiatives, stakeholders, advanced
analytics, data sources, and instrumentation strategies that are
essential to data science success. The book will help you
accelerate and optimize your company's operations through AI and
machine learning. By the end of the book, you will have the tools
and techniques to drive your organization's digital transformation.
Here are a few words from Dr. Kirk Borne, Data Scientist and
Executive Advisor at Booz Allen Hamilton, about the book: Data
analytics should first and foremost be about action and value.
Consequently, the great value of this book is that it seeks to be
actionable. It offers a dynamic progression of purpose-driven
ignition points that you can act upon. What you will learn Train
your organization to transition from being data-driven to being
value-driven Navigate and master the big data business model
maturity index Learn a methodology for determining the economic
value of your data and analytics Understand how AI and machine
learning can create analytics assets that appreciate in value the
more that they are used Become aware of digital transformation
misconceptions and pitfalls Create empowered and dynamic teams that
fuel your organization's digital transformation Who this book is
forThis book is designed to benefit everyone from students who
aspire to study the economic fundamentals behind data and digital
transformation to established business leaders and professionals
who want to learn how to leverage data and analytics to accelerate
their business careers.
Vehicular traffic congestion and accidents remain universal issues
in today's world. Due to the continued growth in the use of
vehicles, optimizing traffic management operations is an immense
challenge. To reduce the number of traffic accidents, improve the
performance of transportation systems, enhance road safety, and
protect the environment, vehicular ad-hoc networks have been
introduced. Current developments in wireless communication,
computing paradigms, big data, and cloud computing enable the
enhancement of these networks, equipped with wireless communication
capabilities and high-performance processing tools. Cloud-Based Big
Data Analytics in Vehicular Ad-Hoc Networks is a pivotal reference
source that provides vital research on cloud and data analytic
applications in intelligent transportation systems. While
highlighting topics such as location routing, accident detection,
and data warehousing, this publication addresses future challenges
in vehicular ad-hoc networks and presents viable solutions. This
book is ideally designed for researchers, computer scientists,
engineers, automobile industry professionals, IT practitioners,
academicians, and students seeking current research on cloud
computing models in vehicular networks.
One-stop solution for NLP practitioners, ML developers, and data
scientists to build effective NLP systems that can perform
real-world complicated tasks Key Features Apply deep learning
algorithms and techniques such as BiLSTMS, CRFs, BPE and more using
TensorFlow 2 Explore applications like text generation,
summarization, weakly supervised labelling and more Read cutting
edge material with seminal papers provided in the GitHub repository
with full working code Book DescriptionRecently, there have been
tremendous advances in NLP, and we are now moving from research
labs into practical applications. This book comes with a perfect
blend of both the theoretical and practical aspects of trending and
complex NLP techniques. The book is focused on innovative
applications in the field of NLP, language generation, and dialogue
systems. It helps you apply the concepts of pre-processing text
using techniques such as tokenization, parts of speech tagging, and
lemmatization using popular libraries such as Stanford NLP and
SpaCy. You will build Named Entity Recognition (NER) from scratch
using Conditional Random Fields and Viterbi Decoding on top of
RNNs. The book covers key emerging areas such as generating text
for use in sentence completion and text summarization, bridging
images and text by generating captions for images, and managing
dialogue aspects of chatbots. You will learn how to apply transfer
learning and fine-tuning using TensorFlow 2. Further, it covers
practical techniques that can simplify the labelling of textual
data. The book also has a working code that is adaptable to your
use cases for each tech piece. By the end of the book, you will
have an advanced knowledge of the tools, techniques and deep
learning architecture used to solve complex NLP problems. What you
will learn Grasp important pre-steps in building NLP applications
like POS tagging Use transfer and weakly supervised learning using
libraries like Snorkel Do sentiment analysis using BERT Apply
encoder-decoder NN architectures and beam search for summarizing
texts Use Transformer models with attention to bring images and
text together Build apps that generate captions and answer
questions about images using custom Transformers Use advanced
TensorFlow techniques like learning rate annealing, custom layers,
and custom loss functions to build the latest DeepNLP models Who
this book is forThis is not an introductory book and assumes the
reader is familiar with basics of NLP and has fundamental Python
skills, as well as basic knowledge of machine learning and
undergraduate-level calculus and linear algebra. The readers who
can benefit the most from this book include intermediate ML
developers who are familiar with the basics of supervised learning
and deep learning techniques and professionals who already use
TensorFlow/Python for purposes such as data science, ML, research,
analysis, etc.
A well thought out, fit-for-purpose data strategy is vital to
modern data-driven businesses. This book is your essential guide to
planning, developing and implementing such a strategy, presenting a
framework which takes you from data strategy definition to
successful strategy delivery and execution with support and
engagement from stakeholders. Key topics include data-driven
business transformation, change enablers, benefits realisation and
measurement.
Learn Azure's cloud capabilities with the help of this introductory
guide to employing Azure for your cloud infrastructure needs. Key
Features Get a clear overview of Azure's capabilities and benefits,
and learn how to get started efficiently Develop the ability to opt
for cloud architecture and design that best fits your organization
Leverage Azure opportunities for cost savings and optimization Book
DescriptionMicrosoft Azure is a powerful cloud computing platform
that offers a multitude of services and capabilities for
organizations of any size moving to a cloud strategy. Azure
Strategy and Implementation Guide Third Edition encapsulates the
entire spectrum of measures involved in Azure deployment that
includes understanding Azure fundamentals, choosing a suitable
cloud architecture, building on design principles, becoming
familiar with Azure DevOps, and learning best practices for
optimization and management. The book begins by introducing you to
the Azure cloud platform and demonstrating the substantial scope of
digital transformation and innovation that can be achieved by
leveraging Azure's capabilities. The guide further acquaints you
with practical insights on application modernization, Azure
Infrastructure as a Service (IaaS) deployment, infrastructure
management, key application architectures, best practices of Azure
DevOps, and Azure automation. By the end of this book, you will be
proficient in driving Azure operations right from the planning and
cloud migration stage to cost management and troubleshooting. What
you will learn Deploy and run Azure infrastructure services Carry
out detailed planning for migrating applications to the cloud with
Azure Move underlying code class structure into a serverless model
Use a gateway to isolate your services and applications Define
roles and responsibilities in DevOps Implement release &
deployment coordination and automation Who this book is forAzure
Strategy and Implementation Guide Third Edition is designed to
benefit Azure architects, cloud solution architects, Azure
developers, Azure administrators, and anyone who wants to develop
an expertise in operating and administering the Azure cloud. A
basic familiarity with operating systems and databases will help
you grasp the concepts covered in this book.
Kickstart your NLP journey by exploring BERT and its variants such
as ALBERT, RoBERTa, DistilBERT, VideoBERT, and more with Hugging
Face's transformers library Key Features Explore the encoder and
decoder of the transformer model Become well-versed with BERT along
with ALBERT, RoBERTa, and DistilBERT Discover how to pre-train and
fine-tune BERT models for several NLP tasks Book DescriptionBERT
(bidirectional encoder representations from transformer) has
revolutionized the world of natural language processing (NLP) with
promising results. This book is an introductory guide that will
help you get to grips with Google's BERT architecture. With a
detailed explanation of the transformer architecture, this book
will help you understand how the transformer's encoder and decoder
work. You'll explore the BERT architecture by learning how the BERT
model is pre-trained and how to use pre-trained BERT for downstream
tasks by fine-tuning it for NLP tasks such as sentiment analysis
and text summarization with the Hugging Face transformers library.
As you advance, you'll learn about different variants of BERT such
as ALBERT, RoBERTa, and ELECTRA, and look at SpanBERT, which is
used for NLP tasks like question answering. You'll also cover
simpler and faster BERT variants based on knowledge distillation
such as DistilBERT and TinyBERT. The book takes you through MBERT,
XLM, and XLM-R in detail and then introduces you to sentence-BERT,
which is used for obtaining sentence representation. Finally,
you'll discover domain-specific BERT models such as BioBERT and
ClinicalBERT, and discover an interesting variant called VideoBERT.
By the end of this BERT book, you'll be well-versed with using BERT
and its variants for performing practical NLP tasks. What you will
learn Understand the transformer model from the ground up Find out
how BERT works and pre-train it using masked language model (MLM)
and next sentence prediction (NSP) tasks Get hands-on with BERT by
learning to generate contextual word and sentence embeddings
Fine-tune BERT for downstream tasks Get to grips with ALBERT,
RoBERTa, ELECTRA, and SpanBERT models Get the hang of the BERT
models based on knowledge distillation Understand cross-lingual
models such as XLM and XLM-R Explore Sentence-BERT, VideoBERT, and
BART Who this book is forThis book is for NLP professionals and
data scientists looking to simplify NLP tasks to enable efficient
language understanding using BERT. A basic understanding of NLP
concepts and deep learning is required to get the best out of this
book.
Learn to extract actionable insights from your big data in real
time using a range of Microsoft Azure features Key Features Updated
with the latest features and new additions to Microsoft Azure
Master the fundamentals of cloud analytics using Azure Learn to use
Azure Synapse Analytics (formerly known as Azure SQL Data
Warehouse) to derive real-time customer insights Book
DescriptionCloud Analytics with Microsoft Azure serves as a
comprehensive guide for big data analysis and processing using a
range of Microsoft Azure features. This book covers everything you
need to build your own data warehouse and learn numerous techniques
to gain useful insights by analyzing big data The book begins by
introducing you to the power of data with big data analytics, the
Internet of Things (IoT), machine learning, artificial
intelligence, and DataOps. You will learn about cloud-scale
analytics and the services Microsoft Azure offers to empower
businesses to discover insights. You will also be introduced to the
new features and functionalities added to the modern data
warehouse. Finally, you will look at two real-world business use
cases to demonstrate high-level solutions using Microsoft Azure.
The aim of these use cases will be to illustrate how real-time data
can be analyzed in Azure to derive meaningful insights and make
business decisions. You will learn to build an end-to-end analytics
pipeline on the cloud with machine learning and deep learning
concepts. By the end of this book, you will be proficient in
analyzing large amounts of data with Azure and using it effectively
to benefit your organization. What you will learn Explore the
concepts of modern data warehouses and data pipelines Discover
unique design considerations while applying a cloud analytics
solution Design an end-to-end analytics pipeline on the cloud
Differentiate between structured, semi-structured, and unstructured
data Choose a cloud-based service for your data analytics solutions
Use Azure services to ingest, store, and analyze data of any scale
Who this book is forThis book is designed to benefit software
engineers, Azure developers, cloud consultants, and anyone who is
keen to learn the process of deriving business insights from huge
amounts of data using Azure. Though not necessary, a basic
understanding of data analytics concepts such as data streaming,
data types, the machine learning life cycle, and Docker containers
will help you get the most out of the book.
Design, secure, and protect the privacy of edge analytics
applications using platforms and tools such as Microsoft's Azure
IoT Edge, MicroPython, and Open Source Computer Vision (OpenCV) Key
Features Become well-versed with best practices for implementing
automated analytical computations Discover real-world examples to
extend cloud intelligence Develop your skills by understanding edge
analytics and applying it to research activities Book
DescriptionEdge analytics has gained attention as the IoT model for
connected devices rises in popularity. This guide will give you
insights into edge analytics as a data analysis model, and help you
understand why it's gaining momentum. You'll begin with the key
concepts and components used in an edge analytics app. Moving
ahead, you'll delve into communication protocols to understand how
sensors send their data to computers or microcontrollers. Next, the
book will demonstrate how to design modern edge analytics apps that
take advantage of the processing power of modern single-board
computers and microcontrollers. Later, you'll explore Microsoft
Azure IoT Edge, MicroPython, and the OpenCV visual recognition
library. As you progress, you'll cover techniques for processing AI
functionalities from the server side to the sensory side of IoT.
You'll even get hands-on with designing a smart doorbell system
using the technologies you've learned. To remove vulnerabilities in
the overall edge analytics architecture, you'll discover ways to
overcome security and privacy challenges. Finally, you'll use tools
to audit and perform real-time monitoring of incoming data and
generate alerts for the infrastructure. By the end of this book,
you'll have learned how to use edge analytics programming
techniques and be able to implement automated analytical
computations. What you will learn Discover the key concepts and
architectures used with edge analytics Understand how to use
long-distance communication protocols for edge analytics Deploy
Microsoft Azure IoT Edge to a Raspberry Pi Create Node-RED
dashboards with MQTT and Text to Speech (TTS) Use MicroPython for
developing edge analytics apps Explore various machine learning
techniques and discover how machine learning is related to edge
analytics Use camera and vision recognition algorithms on the
sensory side to design an edge analytics app Monitor and audit edge
analytics apps Who this book is forIf you are a data analyst, data
architect, or data scientist who is interested in learning and
practicing advanced automated analytical computations, then this
book is for you. You will also find this book useful if you're
looking to learn edge analytics from scratch. Basic knowledge of
data analytics concepts is assumed to get the most out of this
book.
Solve real-world data problems and create data-driven workflows for
easy data movement and processing at scale with Azure Data Factory
Key Features Learn how to load and transform data from various
sources, both on-premises and on cloud Use Azure Data Factory's
visual environment to build and manage hybrid ETL pipelines
Discover how to prepare, transform, process, and enrich data to
generate key insights Book DescriptionAzure Data Factory (ADF) is a
modern data integration tool available on Microsoft Azure. This
Azure Data Factory Cookbook helps you get up and running by showing
you how to create and execute your first job in ADF. You'll learn
how to branch and chain activities, create custom activities, and
schedule pipelines. This book will help you to discover the
benefits of cloud data warehousing, Azure Synapse Analytics, and
Azure Data Lake Gen2 Storage, which are frequently used for big
data analytics. With practical recipes, you'll learn how to
actively engage with analytical tools from Azure Data Services and
leverage your on-premise infrastructure with cloud-native tools to
get relevant business insights. As you advance, you'll be able to
integrate the most commonly used Azure Services into ADF and
understand how Azure services can be useful in designing ETL
pipelines. The book will take you through the common errors that
you may encounter while working with ADF and show you how to use
the Azure portal to monitor pipelines. You'll also understand error
messages and resolve problems in connectors and data flows with the
debugging capabilities of ADF. By the end of this book, you'll be
able to use ADF as the main ETL and orchestration tool for your
data warehouse or data platform projects. What you will learn
Create an orchestration and transformation job in ADF Develop,
execute, and monitor data flows using Azure Synapse Create big data
pipelines using Azure Data Lake and ADF Build a machine learning
app with Apache Spark and ADF Migrate on-premises SSIS jobs to ADF
Integrate ADF with commonly used Azure services such as Azure ML,
Azure Logic Apps, and Azure Functions Run big data compute jobs
within HDInsight and Azure Databricks Copy data from AWS S3 and
Google Cloud Storage to Azure Storage using ADF's built-in
connectors Who this book is forThis book is for ETL developers,
data warehouse and ETL architects, software professionals, and
anyone who wants to learn about the common and not-so-common
challenges faced while developing traditional and hybrid ETL
solutions using Microsoft's Azure Data Factory. You'll also find
this book useful if you are looking for recipes to improve or
enhance your existing ETL pipelines. Basic knowledge of data
warehousing is expected.
People have described nature since the beginning of human history.
They do it for various purposes, including to communicate about
economic, social, governmental, meteorological,
sustainability-related, strategic, military, and survival issues as
well as artistic expression. As a part of the whole world of living
beings, we use various types of senses, known and unknown, labeled
and not identified, to both communicate and create. Describing
Nature Through Visual Data is a collection of impactful research
that discusses issues related to the visualization of scientific
concepts, picturing processes, and products, as well as the role of
computing in advancing visual literacy skills. Organized into four
sections, the book contains descriptions, theories, and examples of
visual and music-based solutions concerning the selected natural or
technological events that are shaping present-day reality. The
chapters pertain to selected scientific fields, digital art,
computer graphics, and new media and confer the possible ways that
visuals, visualization, simulation, and interactive knowledge
presentation can help us to understand and share the content of
scientific thought, research, artistic works, and practice.
Featuring coverage on topics that include mathematical thinking,
music theory, and visual communication, this reference is ideal for
instructors, professionals, researchers, and students keen on
comprehending and enhancing the role of knowledge visualization in
computing, sciences, design, media communication, film,
advertising, and marketing.
Use Microsoft SQL Server 2019 to implement, administer, and secure
a robust database solution that is disaster-proof and highly
available Key Features Explore new features of SQL Server 2019 to
set up, administer, and maintain your database solution
successfully Develop a dynamic SQL Server environment and
streamline big data pipelines Discover best practices for fixing
performance issues, database access management, replication, and
security Book DescriptionSQL Server is one of the most popular
relational database management systems developed by Microsoft. This
second edition of the SQL Server Administrator's Guide will not
only teach you how to administer an enterprise database, but also
help you become proficient at managing and keeping the database
available, secure, and stable. You'll start by learning how to set
up your SQL Server and configure new and existing environments for
optimal use. The book then takes you through designing aspects and
delves into performance tuning by showing you how to use indexes
effectively. You'll understand certain choices that need to be made
about backups, implement security policy, and discover how to keep
your environment healthy. Tools available for monitoring and
managing a SQL Server database, including automating health
reviews, performance checks, and much more, will also be discussed
in detail. As you advance, the book covers essential topics such as
migration, upgrading, and consolidation, along with the techniques
that will help you when things go wrong. Once you've got to grips
with integration with Azure and streamlining big data pipelines,
you'll learn best practices from industry experts for maintaining a
highly reliable database solution. Whether you are an administrator
or are looking to get started with database administration, this
SQL Server book will help you develop the skills you need to
successfully create, design, and deploy database solutions. What
you will learn Discover SQL Server 2019's new features and how to
implement them Fix performance issues by optimizing queries and
making use of indexes Design and use an optimal database management
strategy Combine SQL Server 2019 with Azure and manage your
solution using various automation techniques Implement efficient
backup and recovery techniques in line with security policies Get
to grips with migrating, upgrading, and consolidating with SQL
Server Set up an AlwaysOn-enabled stable and fast SQL Server 2019
environment Understand how to work with Big Data on SQL Server
environments Who this book is forThis book is for database
administrators, database developers, and anyone who wants to
administer large and multiple databases single-handedly using
Microsoft's SQL Server 2019. Basic awareness of database concepts
and experience with previous SQL Server versions is required.
|
|