|
|
Showing 1 - 3 of
3 matches in All Departments
Today, the world is trying to create and educate data scientists
because of the phenomenon of Big Data. And everyone is looking
deeply into this technology. But no one is looking at the larger
architectural picture of how Big Data needs to fit within the
existing systems (data warehousing systems). Taking a look at the
larger picture into which Big Data fits gives the data scientist
the necessary context for how pieces of the puzzle should fit
together. Most references on Big Data look at only one tiny part of
a much larger whole. Until data gathered can be put into an
existing framework or architecture it can't be used to its full
potential. Data Architecture a Primer for the Data Scientist
addresses the larger architectural picture of how Big Data fits
with the existing information infrastructure, an essential topic
for the data scientist. Drawing upon years of practical experience
and using numerous examples and an easy to understand framework.
W.H. Inmon, and Daniel Linstedt define the importance of data
architecture and how it can be used effectively to harness big data
within existing systems. You'll be able to: Turn textual
information into a form that can be analyzed by standard tools.
Make the connection between analytics and Big Data Understand how
Big Data fits within an existing systems environment Conduct
analytics on repetitive and non-repetitive data
Over the past 5 years, the concept of big data has matured, data
science has grown exponentially, and data architecture has become a
standard part of organizational decision-making. Throughout all
this change, the basic principles that shape the architecture of
data have remained the same. There remains a need for people to
take a look at the "bigger picture" and to understand where their
data fit into the grand scheme of things. Data Architecture: A
Primer for the Data Scientist, Second Edition addresses the larger
architectural picture of how big data fits within the existing
information infrastructure or data warehousing systems. This is an
essential topic not only for data scientists, analysts, and
managers but also for researchers and engineers who increasingly
need to deal with large and complex sets of data. Until data are
gathered and can be placed into an existing framework or
architecture, they cannot be used to their full potential. Drawing
upon years of practical experience and using numerous examples and
case studies from across various industries, the authors seek to
explain this larger picture into which big data fits, giving data
scientists the necessary context for how pieces of the puzzle
should fit together.
The Data Vault was invented by Dan Linstedt at the U.S. Department
of Defense, and the standard has been successfully applied to data
warehousing projects at organizations of different sizes, from
small to large-size corporations. Due to its simplified design,
which is adapted from nature, the Data Vault 2.0 standard helps
prevent typical data warehousing failures. "Building a Scalable
Data Warehouse" covers everything one needs to know to create a
scalable data warehouse end to end, including a presentation of the
Data Vault modeling technique, which provides the foundations to
create a technical data warehouse layer. The book discusses how to
build the data warehouse incrementally using the agile Data Vault
2.0 methodology. In addition, readers will learn how to create the
input layer (the stage layer) and the presentation layer (data
mart) of the Data Vault 2.0 architecture including implementation
best practices. Drawing upon years of practical experience and
using numerous examples and an easy to understand framework, Dan
Linstedt and Michael Olschimke discuss: How to load each layer
using SQL Server Integration Services (SSIS), including automation
of the Data Vault loading processes. Important data warehouse
technologies and practices. Data Quality Services (DQS) and Master
Data Services (MDS) in the context of the Data Vault architecture.
|
|