Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > Applications of computing > Databases > Data warehousing
Best practices and invaluable advice from world-renowned data warehouse experts In this book, leading data warehouse experts from the Kimball Group share best practices for using the upcoming "Business Intelligence release" of SQL Server, referred to as SQL Server 2008 R2. In this new edition, the authors explain how SQL Server 2008 R2 provides a collection of powerful new tools that extend the power of its BI toolset to Excel and SharePoint users and they show how to use SQL Server to build a successful data warehouse that supports the business intelligence requirements that are common to most organizations. Covering the complete suite of data warehousing and BI tools that are part of SQL Server 2008 R2, as well as Microsoft Office, the authors walk you through a full project lifecycle, including design, development, deployment and maintenance.Features more than 50 percent new and revised material that covers the rich new feature set of the SQL Server 2008 R2 release, as well as the Office 2010 releaseIncludes brand new content that focuses on PowerPivot for Excel and SharePoint, Master Data Services, and discusses updated capabilities of SQL Server Analysis, Integration, and Reporting ServicesShares detailed case examples that clearly illustrate how to best apply the techniques described in the bookThe accompanying Web site contains all code samples as well as the sample database used throughout the case studies "The Microsoft Data Warehouse Toolkit, Second Edition" provides you with the knowledge of how and when to use BI tools such as Analysis Services and Integration Services to accomplish your most essential data warehousing tasks.
Organization of data warehouses is a vital, but often neglected, aspect of growing an enterprise. Unlike most books on the subject that focus on either the technical aspects of building data warehouses or on business strategies, this valuable reference synthesizes technological know-how with managerial best practices to show how improved alignment between data warehouse plans and business strategies can lead to successful data warehouse adoption capable of supporting an enterprise s entire infrastructure. Strategic Data Warehousing: Achieving Alignment with Business provides data warehouse developers, business managers, and IT professionals and administrators with an integrated approach to achieving successful and sustainable alignment of data warehouses and business goals. More complete than any other text in the field, this comprehensive reference details the joint roles and responsibilities of the data warehouse and business managers in achieving strategic alignment, business user satisfaction, technical integration, and improved flexibility. Complete with case studies that depict real-world scenarios, the text:
Achieving sustainable alignment between the data warehouse and business strategies is a continuous process. Armed with this valuable reference, readers will be able to gain the solid understanding of the organizational, technical, data, and user factors needed to promote a successful data warehouse adoption and become active partners in leveraging this powerful, but often overlooked, information resource.
Discover how graph databases can help you manage and query highly connected data. With this practical book, you'll learn how to design and implement a graph database that brings the power of graphs to bear on a broad range of problem domains. Whether you want to speed up your response to user queries or build a database that can adapt as your business evolves, this book shows you how to apply the schema-free graph model to real-world problems. This second edition includes new code samples and diagrams, using the latest Neo4j syntax, as well as information on new functionality. Learn how different organizations are using graph databases to outperform their competitors. With this book's data modeling, query, and code examples, you'll quickly be able to implement your own solution. Model data with the Cypher query language and property graph model Learn best practices and common pitfalls when modeling with graphs Plan and implement a graph database solution in test-driven fashion Explore real-world examples to learn how and why organizations use a graph database Understand common patterns and components of graph database architecture Use analytical techniques and algorithms to mine graph database information
Cutting-edge content and guidance from a data warehousing expert--now expanded to reflect field trends Data warehousing has revolutionized the way businesses in a wide variety of industries perform analysis and make strategic decisions. Since the first edition of "Data Warehousing Fundamentals," numerous enterprises have implemented data warehouse systems and reaped enormous benefits. Many more are in the process of doing so. Now, this new, revised edition covers the essential fundamentals of data warehousing and business intelligence as well as significant recent trends in the field. The author provides an enhanced, comprehensive overview of data warehousing together with in-depth explanations of critical issues in planning, design, deployment, and ongoing maintenance. IT professionals eager to get into the field will gain a clear understanding of techniques for data extraction from source systems, data cleansing, data transformations, data warehouse architecture and infrastructure, and the various methods for information delivery. This practical "Second Edition" highlights the areas of data warehousing and business intelligence where high-impact technological progress has been made. Discussions on developments include data marts, real-time information delivery, data visualization, requirements gathering methods, multi-tier architecture, OLAP applications, Web clickstream analysis, data warehouse appliances, and data mining techniques. The book also contains review questions and exercises for each chapter, appropriate for self-study or classroom work, industry examples of real-world situations, and several appendices with valuable information. Specifically written for professionals responsible for designing, implementing, or maintaining data warehousing systems, "Data Warehousing Fundamentals" presents agile, thorough, and systematic development principles for the IT professional and anyone working or researching in information management.
Develop the must-have skills required for any data scientist to get the best results from Azure Databricks. Key Features * Learn to develop and productionize ML pipelines using the Databricks Unified Analytics platform * See how to use AutoML, Feature Stores, and MLOps with Databricks * Get a complete understanding of data governance and model deployment Book Description In this book, you'll get to grips with Databricks, enabling you to power-up your organization's data science applications. We'll walk through applying the Databricks AI and ML stack to real-world use cases for natural language processing, computer vision, time series data, and more. We'll dive deep into the complete model development life cycle for data ingestion and analysis, and get familiar with the latest offerings of AutoML, Feature Store, and MLStudio, on the Databricks platform. You'll get hands-on experience implementing repeatable ML operations (MLOps) pipeline using MLFlow, track model training and key metrics, and explore real-time ML, anomaly detection, and streaming analytics with Delta lake and Spark Structured Streaming. Starting with an overview of Data Science use cases across different organizations and industries, you will then be introduced to feature stores, feature tables, and how to access them. You will see why AutoML is important and how to create a baseline model with AutoML within Databricks. Utilizing the ML Flow model registry to manage model versioning and transition to production will be covered, along with detecting and protecting against model drift in production environments. By the end of the book, you will know how to set up your Databricks ML development and deployment as a CI/CD pipeline. What you will learn * Perform natural language processing, computer vision, and more * Explore AutoML, Feature Store, and MLStudio on Databricks * Dive deep into the complete model development life cycle * Experience implementing repeatable MLOps pipelines using MLFlow * Track model training and key metrics * Explore real-time ML, anomaly detection, and streaming analytics * Learn how to handle model drift Who This Book Is For In this book we are going to specifically focus on the tools catering to the Data Scientist persona. Readers who want to learn how to successfully build and deploy end-end Data Science projects using the Databricks cloud agnostic unified analytics platform will benefit from this book, along with AI and Machine Learning practitioners.
Corporations and governmental agencies of all sizes are embracing a new generation of enterprise-scale business intelligence (BI) and data warehousing (DW), and very often appoint a single senior-level individual to serve as the Enterprise BI/DW Program Manager. This book is the essential guide to the incremental and iterative build-out of a successful enterprise-scale BI/DW program comprised of multiple underlying projects, and what the Enterprise Program Manager must successfully accomplish to orchestrate the many moving parts in the quest for true enterprise-scale business intelligence and data warehousing. Author Alan Simon has served as an enterprise business intelligence and data warehousing program management advisor to many of his clients, and spent an entire year with a single client as the adjunct consulting director for a $10 million enterprise data warehousing (EDW) initiative. He brings a wealth of knowledge about best practices, risk management, organizational culture alignment, and other Critical Success Factors (CSFs) to the discipline of enterprise-scale business intelligence and data warehousing.
"Data Warehousing in the Age of the Big Data "will help you and your organization make the most of unstructured data with your existing data warehouse. As Big Data continues to revolutionize how we use data, it doesn't have to create more confusion. Expert author Krish Krishnan helps you make sense of how Big Data fits into the world of data warehousing in clear and concise detail. The book is presented in three distinct parts. Part 1 discusses Big Data, its technologies and use cases from early adopters. Part 2 addresses data warehousing, its shortcomings, and new architecture options, workloads, and integration techniques for Big Data and the data warehouse. Part 3 deals with data governance, data visualization, information life-cycle management, data scientists, and implementing a Big Data-ready data warehouse. Extensive appendixes include case studies from vendor implementations and a special segment on how we can build a healthcare information factory. Ultimately, this book will help you navigate through the complex
layers of Big Data and data warehousing while providing you
information on how to effectively think about using all these
technologies and the architectures to design the next-generation
data warehouse.
This two-volume set constitutes the refereed proceedings of the 17th International Conference on Collaborative Computing: Networking, Applications, and Worksharing, CollaborateCom 2021, held in October 2021. Due to COVID-19 pandemic the conference was held virtually.The 62 full papers and 7 short papers presented were carefully reviewed and selected from 206 submissions. The papers reflect the conference sessions as follows: Optimization for Collaborate System; Optimization based on Collaborative Computing; UVA and Traffic system; Recommendation System; Recommendation System & Network and Security; Network and Security; Network and Security & IoT and Social Networks; IoT and Social Networks & Images handling and human recognition; Images handling and human recognition & Edge Computing; Edge Computing; Edge Computing & Collaborative working; Collaborative working & Deep Learning and application; Deep Learning and application; Deep Learning and application; Deep Learning and application & UVA.
Building upon his earlier book that detailed agile data warehousing programming techniques for the Scrum master, Ralph's latest work illustrates the agile interpretations of the remaining software engineering disciplines: Requirements management benefits from streamlined templates that not only define projects quickly, but ensure nothing essential is overlooked. Data engineering receives two new "hyper modeling" techniques, yielding data warehouses that can be easily adapted when requirements change without having to invest in ruinously expensive data-conversion programs. Quality assurance advances with not only a stereoscopic top-down and bottom-up planning method, but also the incorporation of the latest in automated test engines. Use this step-by-step guide to deepen your own application development skills through self-study, show your teammates the world's fastest and most reliable techniques for creating business intelligence systems, or ensure that the IT department working for you is building your next decision support system the right way.
An easy-to-follow introduction to support vector machines This book provides an in-depth, easy-to-follow introduction to support vector machines drawing only from minimal, carefully motivated technical and mathematical background material. It begins with a cohesive discussion of machine learning and goes on to cover: Knowledge discovery environments Describing data mathematically Linear decision surfaces and functions Perceptron learning Maximum margin classifiers Support vector machines Elements of statistical learning theory Multi-class classification Regression with support vector machines Novelty detection Complemented with hands-on exercises, algorithm descriptions, and data sets, Knowledge Discovery with Support Vector Machines is an invaluable textbook for advanced undergraduate and graduate courses. It is also an excellent tutorial on support vector machines for professionals who are pursuing research in machine learning and related areas.
This book provides readers the "big picture" and a comprehensive survey of the domain of big data processing systems. For the past decade, the Hadoop framework has dominated the world of big data processing, yet recently academia and industry have started to recognize its limitations in several application domains and thus, it is now gradually being replaced by a collection of engines that are dedicated to specific verticals (e.g. structured data, graph data, and streaming data). The book explores this new wave of systems, which it refers to as Big Data 2.0 processing systems. After Chapter 1 presents the general background of the big data phenomena, Chapter 2 provides an overview of various general-purpose big data processing systems that allow their users to develop various big data processing jobs for different application domains. In turn, Chapter 3 examines various systems that have been introduced to support the SQL flavor on top of the Hadoop infrastructure and provide competing and scalable performance in the processing of large-scale structured data. Chapter 4 discusses several systems that have been designed to tackle the problem of large-scale graph processing, while the main focus of Chapter 5 is on several systems that have been designed to provide scalable solutions for processing big data streams, and on other sets of systems that have been introduced to support the development of data pipelines between various types of big data processing jobs and systems. Next, Chapter 6 focuses on covering the emerging frameworks and systems in the domain of scalable machine learning and deep learning processing. Lastly, Chapter 7 shares conclusions and an outlook on future research challenges. This new and considerably enlarged second edition not only contains the completely new chapter 6, but also offers a refreshed content for the state-of-the-art in all domains of big data processing over the last years. Overall, the book offers a valuable reference guide for professional, students, and researchers in the domain of big data processing systems. Further, its comprehensive content will hopefully encourage readers to pursue further research on the subject.
This guide shows how to combine data science with social science to gain unprecedented insight into customer behavior, so you can change it. Joanne Rodrigues-Craig bridges the gap between predictive data science and statistical techniques that reveal why important things happen -- why customers buy more, or why they immediately leave your site -- so you can get more behaviors you want and less you don't. Drawing on extensive enterprise experience and deep knowledge of demographics and sociology, Rodrigues-Craig shows how to create better theories and metrics, so you can accelerate the process of gaining insight, altering behavior, and earning business value. You'll learn how to: Develop complex, testable theories for understanding individual and social behavior in web products Think like a social scientist and contextualize individual behavior in today's social environments Build more effective metrics and KPIs for any web product or system Conduct more informative and actionable A/B tests Explore causal effects, reflecting a deeper understanding of the differences between correlation and causation Alter user behavior in a complex web product Understand how relevant human behaviors develop, and the prerequisites for changing them Choose the right statistical techniques for common tasks such as multistate and uplift modeling Use advanced statistical techniques to model multidimensional systems Do all of this in R (with sample code available in a separate code manual)
Do your business intelligence (BI) projects take too long to deliver? Is the value of the deliverables less than satisfactory? Do these projects propagate poor data management practices? If you screamed yes to any of these questions, read this book to master a proven approach to building your enterprise data warehouse and BI initiatives. "Extreme Scoping", based on the Business Intelligence Roadmap, will show you how to build analytics applications rapidly yet not sacrifice data management and enterprise architecture. In addition, all of the roles required to deliver all seven steps of this agile methodology are explained along with many real-world examples. From Wayne Eckersons Foreword -- I've read many books about data warehousing and business intelligence (BI). This book by Larissa Moss is one of the best. I should not be surprised. Larissa has spent years refining the craft of designing, building, and delivering BI applications. Over the years, she has developed a keen insight about what works and doesnt work in BI. This book brings to light the wealth of that development experience. Best of all, this is not some dry text that laboriously steps readers through a technical methodology. Larissa expresses her ideas in a clear, concise, and persuasive manner. I highlighted so many beautifully written and insightful paragraphs in her manuscript that it became comical. I desperately wanted the final, published book rather than the manuscript so I could dog-ear it to death and place it front-and-center in my office bookshelf! From David Wells Foreword : Extreme Scoping is rich with advice and guidance for virtually every aspect of BI projects from planning and requirements to deployment and from back-end data management to front-end information and analytics services. Larissa is both a pragmatist and an independent thinker. Those qualities come through in the style of this book. This is a well-written book that is easy to absorb. It is not full of surprises. It is filled with a lot of common sense and lessons learned through experience.
The chapter "An Efficient Index for Reachability Queries in Public Transport Networks" is available open access under a Creative Commons Attribution 4.0 International License via link.springer.com.
Jede Business-Intelligence-Anwendung beruht letzten Endes auf einem Data Warehouse. Data Warehousing ist deshalb ein sehr wichtiges Gebiet der Angewandten Informatik, insbesondere im Zeitalter von Big Data. Das vorliegende Buch beleuchtet das Data Warehouse aus zwei Perspektiven: der des Entwicklers und der des Anwenders. Der zukA1/4nftige Entwickler lernt, ein Data Warehouse mit geeigneten Methoden selbst zu entwickeln. FA1/4r den zukA1/4nftigen Anwender geht der Autor auf die Themen Reporting, Online Analytical Processing und Data Mining ein. Das Lehrbuch ist auch zum Selbststudium geeignet. Kenntnisse A1/4ber Datenbanksysteme sollten allerdings vorhanden sein.
Are you struggling with the formal design of your organisation's data resource? Do you find yourself forced into generic data architectures and universal data models? Do you find yourself warping the business to fit a purchased application? Do you find yourself pushed into developing physical databases without formal logical design? Do you find disparate data throughout the organisation? If the answer to any of these questions is Yes, then you need to read Data Resource Design to help guide you through a formal design process that produces a high quality data resource within a single common data architecture. Most public and private sector organisations do not consistently follow a formal data resource design process that begins with the organisation's perception of the business world, proceeds through logical data design, through physical data design, and into implementation. Most organisations charge ahead with physical database implementation, physical package implementation, and other brute-force-physical approaches. The result is a data resource that becomes disparate and does not fully support the organisation in its business endeavours. This book describes how to formally design an organisation's data resource to meet its current and future business information demand. It builds on "Data Resource Simplexity", which described how to stop the burgeoning data disparity, and on "Data Resource Integration", which described how to understand and resolve an organisation's disparate data resource. It describes the concepts, principles, and techniques for building a high quality data resource based on an organisation's perception of the business world in which they operate. Like "Data Resource Simplexity" and "Data Resource Integration", Michael Brackett draws on five decades of data management experience building and managing data resources, and resolving disparate data in both public and private sector organisations. He leverages theories, concepts, principles, and techniques from a wide variety of disciplines, such as human dynamics, mathematics, physics, chemistry, philosophy, and biology, and applies them to properly designing data as a critical resource of an organisation. He shows how to understand the business environment where an organisation operates and design a data resource that supports the organisation in that business environment.
This book offers comprehensive coverage of information retrieval by considering both Text Based Information Retrieval (TBIR) and Content Based Image Retrieval (CBIR), together with new research topics. The approach to TBIR is based on creating a thesaurus, as well as event classification and detection. N-gram thesaurus generation for query refinement offers a new method for improving the precision of retrieval, while event classification and detection approaches aid in the classification and organization of information using web documents for domain-specific retrieval applications. In turn, with regard to content based image retrieval (CBIR) the book presents a histogram construction method, which is based on human visual perceptions of color. The book's overarching goal is to introduce readers to new ideas in an easy-to-follow manner.
This book constitutes the thoroughly refereed post-conference proceedings of the Third COST Action IC1302 International KEYSTONE Conference on Semantic Keyword-Based Search on Structured Data Sources, IKC 2017, held in Gdansk, Poland, in September 2017. The 13 revised full papers and 5 short papers included in the first part of the book were carefully reviewed and selected from numerous submissions. The second part contains reports that summarize the major activities and achievements that have taken place in the context of the action: the short term scientific missions, the outcome of the summer schools, and the results achieved within the following four work packages: representation of structured data sources; keyword search; user interaction and keyword query interpretation; and research integration, showcases, benchmarks and evaluations. Also included is a short report generated by the chairs of the action. The papers cover a broad range of topics in the area of keyword search combining expertise from many different related fields such as information retrieval, natural language processing, ontology management, indexing, semantic web and linked data.
Learn essential techniques from data warehouse legend Bill Inmon on how to build the reporting environment your business needs now! Answers for many valuable business questions hide in text. How well can your existing reporting environment extract the necessary text from email, spreadsheets, and documents, and put it in a useful format for analytics and reporting? Transforming the traditional data warehouse into an efficient unstructured data warehouse requires additional skills from the analyst, architect, designer, and developer. This book will prepare you to successfully implement an unstructured data warehouse and, through clear explanations, examples, and case studies, you will learn new techniques and tips to successfully obtain and analyse text. Master these ten objectives: Build an unstructured data warehouse using the 11-step approach; Integrate text and describe it in terms of homogeneity, relevance, medium, volume, and structure; Overcome challenges including blather, the Tower of Babel, and lack of natural relationships; Avoid the Data Junkyard and combat the "Spiders Web"; Reuse techniques perfected in the traditional data warehouse and Data Warehouse 2.0 , including iterative development; Apply essential techniques for textual Extract, Transform, and Load (ETL) such as phrase recognition, stop word filtering, and synonym replacement; Design the Document Inventory system and link unstructured text to structured data; Leverage indexes for efficient text analysis and taxonomies for useful external categorisation; Manage large volumes of data using advanced techniques such as backward pointers; Evaluate technology choices suitable for unstructured data processing, such as data warehouse appliances.
Until recently, many people thought big data was a passing fad. "Data science" was an enigmatic term. Today, big data is taken seriously, and data science is considered downright sexy. With this anthology of reports from award-winning journalist Mike Barlow, you'll appreciate how data science is fundamentally altering our world, for better and for worse. Barlow paints a picture of the emerging data space in broad strokes. From new techniques and tools to the use of data for social good, you'll find out how far data science reaches. With this anthology, you'll learn how: Analysts can now get results from their data queries in near real time Indie manufacturers are blurring the lines between hardware and software Companies try to balance their desire for rapid innovation with the need to tighten data security Advanced analytics and low-cost sensors are transforming equipment maintenance from a cost center to a profit center CIOs have gradually evolved from order takers to business innovators New analytics tools let businesses go beyond data analysis and straight to decision-making Mike Barlow is an award-winning journalist, author, and communications strategy consultant. Since launching his own firm, Cumulus Partners, he has represented major organizations in a number of industries.
Most of modern enterprises, institutions, and organizations rely on knowledge-based management systems. In these systems, knowledge is gained from data analysis. Nowadays, knowledge-based management systems include data warehouses as their core components. The purpose of building a data warehouse is twofold. Firstly, to integrate multiple heterogeneous, autonomous, and distributed data sources within an enterprise. Secondly, to provide a platform for advanced, complex, and efficient data analysis. Data integrated in a data warehouse are analyzed by the so-called On-Line Analytical Processing (OLAP) applications designed among others for discovering trends, patterns of behavior, and anomalies as well as for finding dependencies between data. Massive amounts of integrated data and the complexity of integrated data that more and more often come from WEB-based, XML-based, spatio-temporal, object, and multimedia systems, make data integration and processing challenging. The objective of NEW TRENDS IN DATA WAREHOUSING AND DATA ANALYSIS is fourfold: First, to bring together the most recent research and practical achievements in the DW and OLAP technologies. Second, to open and discuss new, just emerging areas of further development. Third, to provide the up-to-date bibliography of published works and the resource of research achievements for anyone interested in up-to-date data warehouse issues. And, finally, to assist in the dissemination of knowledge in the field of advanced DW and OLAP.
Prepare for Microsoft Exam 70-767-and help demonstrate your real-world mastery of skills for managing data warehouses. This exam is intended for Extract, Transform, Load (ETL) data warehouse developers who create business intelligence (BI) solutions. Their responsibilities include data cleansing as well as ETL and data warehouse implementation. The reader should have experience installing and implementing a Master Data Services (MDS) model, using MDS tools, and creating a Master Data Manager database and web application. The reader should understand how to design and implement ETL control flow elements and work with a SQL Service Integration Services package. Focus on the expertise measured by these objectives: * Design, and implement, and maintain a data warehouse * Extract, transform, and load data * Build data quality solutionsThis Microsoft Exam Ref: * Organizes its coverage by exam objectives * Features strategic, what-if scenarios to challenge you * Assumes you have working knowledge of relational database technology and incremental database extraction, as well as experience with designing ETL control flows, using and debugging SSIS packages, accessing and importing or exporting data from multiple sources, and managing a SQL data warehouse. Implementing a SQL Data Warehouse About the Exam Exam 70-767 focuses on skills and knowledge required for working with relational database technology. About Microsoft Certification Passing this exam earns you credit toward a Microsoft Certified Professional (MCP) or Microsoft Certified Solutions Associate (MCSA) certification that demonstrates your mastery of data warehouse management Passing this exam as well as Exam 70-768 (Developing SQL Data Models) earns you credit toward a Microsoft Certified Solutions Associate (MCSA) SQL 2016 Business Intelligence (BI) Development certification. See full details at: microsoft.com/learning |
You may like...
E-Discovery Tools and Applications in…
Egbert de Smet, Sangeeta Dhamdhere
Hardcover
R5,151
Discovery Miles 51 510
Big Data Management, Technologies, and…
Wen-Chen Hu, Naima Kaabouch
Hardcover
R4,663
Discovery Miles 46 630
Innovations in XML Applications and…
Jose Carlos Ramalho, Alberto Simoes, …
Hardcover
R5,123
Discovery Miles 51 230
Artificial Intelligence Applications and…
Ilias Maglogiannis, Lazaros Iliadis, …
Hardcover
R2,875
Discovery Miles 28 750
Artificial Intelligence on Medical Data…
Mousumi Gupta, Sujata Ghatak, …
Hardcover
R4,291
Discovery Miles 42 910
|