![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Databases > Data capture & analysis
Get to grips with building and productionizing end-to-end big data solutions in Azure and learn best practices for working with large datasets Key Features Integrate with Azure Synapse Analytics, Cosmos DB, and Azure HDInsight Kafka Cluster to scale and analyze your projects and build pipelines Use Databricks SQL to run ad hoc queries on your data lake and create dashboards Productionize a solution using CI/CD for deploying notebooks and Azure Databricks Service to various environments Book DescriptionAzure Databricks is a unified collaborative platform for performing scalable analytics in an interactive environment. The Azure Databricks Cookbook provides recipes to get hands-on with the analytics process, including ingesting data from various batch and streaming sources and building a modern data warehouse. The book starts by teaching you how to create an Azure Databricks instance within the Azure portal, Azure CLI, and ARM templates. You'll work through clusters in Databricks and explore recipes for ingesting data from sources, including files, databases, and streaming sources such as Apache Kafka and EventHub. The book will help you explore all the features supported by Azure Databricks for building powerful end-to-end data pipelines. You'll also find out how to build a modern data warehouse by using Delta tables and Azure Synapse Analytics. Later, you'll learn how to write ad hoc queries and extract meaningful insights from the data lake by creating visualizations and dashboards with Databricks SQL. Finally, you'll deploy and productionize a data pipeline as well as deploy notebooks and Azure Databricks service using continuous integration and continuous delivery (CI/CD). By the end of this Azure book, you'll be able to use Azure Databricks to streamline different processes involved in building data-driven apps. What you will learn Read and write data from and to various Azure resources and file formats Build a modern data warehouse with Delta Tables and Azure Synapse Analytics Explore jobs, stages, and tasks and see how Spark lazy evaluation works Handle concurrent transactions and learn performance optimization in Delta tables Learn Databricks SQL and create real-time dashboards in Databricks SQL Integrate Azure DevOps for version control, deploying, and productionizing solutions with CI/CD pipelines Discover how to use RBAC and ACLs to restrict data access Build end-to-end data processing pipeline for near real-time data analytics Who this book is forThis recipe-based book is for data scientists, data engineers, big data professionals, and machine learning engineers who want to perform data analytics on their applications. Prior experience of working with Apache Spark and Azure is necessary to get the most out of this book.
Think about your data intelligently and ask the right questions Key Features Master data cleaning techniques necessary to perform real-world data science and machine learning tasks Spot common problems with dirty data and develop flexible solutions from first principles Test and refine your newly acquired skills through detailed exercises at the end of each chapter Book DescriptionData cleaning is the all-important first step to successful data science, data analysis, and machine learning. If you work with any kind of data, this book is your go-to resource, arming you with the insights and heuristics experienced data scientists had to learn the hard way. In a light-hearted and engaging exploration of different tools, techniques, and datasets real and fictitious, Python veteran David Mertz teaches you the ins and outs of data preparation and the essential questions you should be asking of every piece of data you work with. Using a mixture of Python, R, and common command-line tools, Cleaning Data for Effective Data Science follows the data cleaning pipeline from start to end, focusing on helping you understand the principles underlying each step of the process. You'll look at data ingestion of a vast range of tabular, hierarchical, and other data formats, impute missing values, detect unreliable data and statistical anomalies, and generate synthetic features. The long-form exercises at the end of each chapter let you get hands-on with the skills you've acquired along the way, also providing a valuable resource for academic courses. What you will learn Ingest and work with common data formats like JSON, CSV, SQL and NoSQL databases, PDF, and binary serialized data structures Understand how and why we use tools such as pandas, SciPy, scikit-learn, Tidyverse, and Bash Apply useful rules and heuristics for assessing data quality and detecting bias, like Benford's law and the 68-95-99.7 rule Identify and handle unreliable data and outliers, examining z-score and other statistical properties Impute sensible values into missing data and use sampling to fix imbalances Use dimensionality reduction, quantization, one-hot encoding, and other feature engineering techniques to draw out patterns in your data Work carefully with time series data, performing de-trending and interpolation Who this book is forThis book is designed to benefit software developers, data scientists, aspiring data scientists, teachers, and students who work with data. If you want to improve your rigor in data hygiene or are looking for a refresher, this book is for you. Basic familiarity with statistics, general concepts in machine learning, knowledge of a programming language (Python or R), and some exposure to data science are helpful.
Dieser Buchtitel ist Teil des Digitalisierungsprojekts Springer Book Archives mit Publikationen, die seit den Anfangen des Verlags von 1842 erschienen sind. Der Verlag stellt mit diesem Archiv Quellen fur die historische wie auch die disziplingeschichtliche Forschung zur Verfugung, die jeweils im historischen Kontext betrachtet werden mussen. Dieser Titel erschien in der Zeit vor 1945 und wird daher in seiner zeittypischen politisch-ideologischen Ausrichtung vom Verlag nicht beworben.
A practical blockchain handbook designed to take you through implementing and re-engineering banking and financial solutions and workflows using eight step-by-step projects Key Features Implement various end-to-end blockchain projects and learn to enhance present-day financial solutions Use Ethereum, Hyperledger, and Stellar to build public and private decentralized applications Address complex challenges faced in the BFSI domain using different blockchain platform services Book DescriptionBlockchain technology will continue to play an integral role in the banking and finance sector in the coming years. It will enable enterprises to build transparent and secure business processes. Experts estimate annual savings of up to 20 billion dollars from this technology. This book will help you build financial apps using blockchain, guiding you through enhancing popular products and services in the banking and finance sector. The book starts by explaining the essential concepts of blockchain, and the impact of blockchain technology on the BFSI sector. Next, you'll delve into re-designing existing banking processes and building new financial apps using blockchain. To accomplish this, you'll work through eight blockchain projects. By demonstrating the entire process, the book helps you understand everything from setting up the environment and building frontend portals to system integration and testing apps. You will gain hands-on experience with the Ethereum, Hyperledger Fabric, and Stellar to develop private and public decentralized apps. Finally, you'll learn how to use ancillary platforms and frameworks such as IPFS, Truffle OpenZeppelin, and MetaMask. By the end of this blockchain book, you'll have an in-depth understanding of how to leverage distributed ledgers and smart contracts for financial use cases. What you will learn Design and implement blockchain solutions in a BFSI organization Explore common architectures and implementation models for enterprise blockchain Design blockchain wallets for multi-purpose applications using Ethereum Build secure and fast decentralized trading ecosystems with Blockchain Implement smart contracts to build secure process workflows in Ethereum and Hyperledger Fabric Use the Stellar platform to build KYC and AML-compliant remittance workflows Map complex business workflows and automate backend processes in a blockchain architecture Who this book is forThis book is for blockchain and Dapps developers, or anyone looking for a guide to building innovative and highly secure solutions in the fintech domain using real-world use cases. Developers working in financial enterprises and banks, and solution architects looking to build brand new process flows using blockchain technology will also find the book useful. Experience with Solidity programming and prior knowledge of finance and trade are required to get the most out of this book.
A beginner's guide to storing, managing, and analyzing data with the updated features of Elastic 7.0 Key Features Gain access to new features and updates introduced in Elastic Stack 7.0 Grasp the fundamentals of Elastic Stack including Elasticsearch, Logstash, and Kibana Explore useful tips for using Elastic Cloud and deploying Elastic Stack in production environments Book DescriptionThe Elastic Stack is a powerful combination of tools for techniques such as distributed search, analytics, logging, and visualization of data. Elastic Stack 7.0 encompasses new features and capabilities that will enable you to find unique insights into analytics using these techniques. This book will give you a fundamental understanding of what the stack is all about, and help you use it efficiently to build powerful real-time data processing applications. The first few sections of the book will help you understand how to set up the stack by installing tools, and exploring their basic configurations. You'll then get up to speed with using Elasticsearch for distributed searching and analytics, Logstash for logging, and Kibana for data visualization. As you work through the book, you will discover the technique of creating custom plugins using Kibana and Beats. This is followed by coverage of the Elastic X-Pack, a useful extension for effective security and monitoring. You'll also find helpful tips on how to use Elastic Cloud and deploy Elastic Stack in production environments. By the end of this book, you'll be well versed with the fundamental Elastic Stack functionalities and the role of each component in the stack to solve different data processing problems. What you will learn Install and configure an Elasticsearch architecture Solve the full-text search problem with Elasticsearch Discover powerful analytics capabilities through aggregations using Elasticsearch Build a data pipeline to transfer data from a variety of sources into Elasticsearch for analysis Create interactive dashboards for effective storytelling with your data using Kibana Learn how to secure, monitor and use Elastic Stack's alerting and reporting capabilities Take applications to an on-premise or cloud-based production environment with Elastic Stack Who this book is forThis book is for entry-level data professionals, software engineers, e-commerce developers, and full-stack developers who want to learn about Elastic Stack and how the real-time processing and search engine works for business analytics and enterprise search applications. Previous experience with Elastic Stack is not required, however knowledge of data warehousing and database concepts will be helpful.
Learn how to gain insights from your data as well as machine learning and become a presentation pro who can create interactive dashboards Key Features Enhance your presentation skills by implementing engaging data storytelling and visualization techniques Learn the basics of machine learning and easily apply machine learning models to your data Improve productivity by automating your data processes Book DescriptionData Analytics Made Easy is an accessible beginner's guide for anyone working with data. The book interweaves four key elements: Data visualizations and storytelling - Tired of people not listening to you and ignoring your results? Don't worry; chapters 7 and 8 show you how to enhance your presentations and engage with your managers and co-workers. Learn to create focused content with a well-structured story behind it to captivate your audience. Automating your data workflows - Improve your productivity by automating your data analysis. This book introduces you to the open-source platform, KNIME Analytics Platform. You'll see how to use this no-code and free-to-use software to create a KNIME workflow of your data processes just by clicking and dragging components. Machine learning - Data Analytics Made Easy describes popular machine learning approaches in a simplified and visual way before implementing these machine learning models using KNIME. You'll not only be able to understand data scientists' machine learning models; you'll be able to challenge them and build your own. Creating interactive dashboards - Follow the book's simple methodology to create professional-looking dashboards using Microsoft Power BI, giving users the capability to slice and dice data and drill down into the results. What you will learn Understand the potential of data and its impact on your business Import, clean, transform, combine data feeds, and automate your processes Influence business decisions by learning to create engaging presentations Build real-world models to improve profitability, create customer segmentation, automate and improve data reporting, and more Create professional-looking and business-centric visuals and dashboards Open the lid on the black box of AI and learn about and implement supervised and unsupervised machine learning models Who this book is forThis book is for beginners who work with data and those who need to know how to interpret their business/customer data. The book also covers the high-level concepts of data workflows, machine learning, data storytelling, and visualizations, which are useful for managers. No previous math, statistics, or computer science knowledge is required.
Master scala's advanced techniques to solve real-world problems in data analysis and gain valuable insights from your data Key Features A beginner's guide for performing data analysis loaded with numerous rich, practical examples Access to popular Scala libraries such as Breeze, Saddle for efficient data manipulation and exploratory analysis Develop applications in Scala for real-time analysis and machine learning in Apache Spark Book DescriptionEfficient business decisions with an accurate sense of business data helps in delivering better performance across products and services. This book helps you to leverage the popular Scala libraries and tools for performing core data analysis tasks with ease. The book begins with a quick overview of the building blocks of a standard data analysis process. You will learn to perform basic tasks like Extraction, Staging, Validation, Cleaning, and Shaping of datasets. You will later deep dive into the data exploration and visualization areas of the data analysis life cycle. You will make use of popular Scala libraries like Saddle, Breeze, Vegas, and PredictionIO for processing your datasets. You will learn statistical methods for deriving meaningful insights from data. You will also learn to create applications for Apache Spark 2.x on complex data analysis, in real-time. You will discover traditional machine learning techniques for doing data analysis. Furthermore, you will also be introduced to neural networks and deep learning from a data analysis standpoint. By the end of this book, you will be capable of handling large sets of structured and unstructured data, perform exploratory analysis, and building efficient Scala applications for discovering and delivering insights What you will learn Techniques to determine the validity and confidence level of data Apply quartiles and n-tiles to datasets to see how data is distributed into many buckets Create data pipelines that combine multiple data lifecycle steps Use built-in features to gain a deeper understanding of the data Apply Lasso regression analysis method to your data Compare Apache Spark API with traditional Apache Spark data analysis Who this book is forIf you are a data scientist or a data analyst who wants to learn how to perform data analysis using Scala, this book is for you. All you need is knowledge of the basic fundamentals of Scala programming.
An expert guide to implementing fast, secure, and scalable decentralized applications that work with thousands of users in real time Key Features Implement advanced features of the Ethereum network to build powerful decentralized applications Build smart contracts on different domains using the programming techniques of Solidity and Vyper Explore the architecture of Ethereum network to understand advanced use cases of blockchain development Book DescriptionEthereum is one of the commonly used platforms for building blockchain applications. It's a decentralized platform for applications that can run exactly as programmed without being affected by fraud, censorship, or third-party interference. This book will give you a deep understanding of how blockchain works so that you can discover the entire ecosystem, core components, and its implementations. You will get started by understanding how to configure and work with various Ethereum protocols for developing dApps. Next, you will learn to code and create powerful smart contracts that scale with Solidity and Vyper. You will then explore the building blocks of the dApps architecture, and gain insights on how to create your own dApp through a variety of real-world examples. The book will even guide you on how to deploy your dApps on multiple Ethereum instances with the required best practices and techniques. The next few chapters will delve into advanced topics such as, building advanced smart contracts and multi-page frontends using Ethereum blockchain. You will also focus on implementing machine learning techniques to build decentralized autonomous applications, in addition to covering several use cases across a variety of domains such as, social media and e-commerce. By the end of this book, you will have the expertise you need to build decentralized autonomous applications confidently. What you will learn Apply scalability solutions on dApps with Plasma and state channels Understand the important metrics of blockchain for analyzing and determining its state Develop a decentralized web application using React.js and Node.js Create oracles with Node.js to provide external data to smart contracts Get to grips with using Etherscan and block explorers for various transactions Explore web3.js, Solidity, and Vyper for dApps communication Deploy apps with multiple Ethereum instances including TestRPC, private chain, test chain, and mainnet Who this book is forThis book is for anyone who wants to build fast, highly secure, and transactional decentralized applications. If you are an Ethereum developer looking to perfect your existing skills in building powerful blockchain applications, then this book is for you. Basic knowledge of Ethereum and blockchain is necessary to understand the concepts covered in this book.
Understand, design, and create cognitive applications using Watson's suite of APIs. Key Features Develop your skills and work with IBM Watson APIs to build efficient and powerful cognitive apps Learn how to build smart apps to carry out different sets of activities using real-world use cases Get well versed with the best practices of IBM Watson and implement them in your daily work Book DescriptionCognitive computing is rapidly infusing every aspect of our lives riding on three important fields: data science, machine learning (ML), and artificial intelligence (AI). It allows computing systems to learn and keep on improving as the amount of data in the system grows. This book introduces readers to a whole new paradigm of computing - a paradigm that is totally different from the conventional computing of the Information Age. You will learn the concepts of ML, deep learning (DL), neural networks, and AI through the set of APIs provided by IBM Watson. This book will help you build your own applications to understand, plan, and solve problems, and analyze them as per your needs. You will learn about various domains of cognitive computing, such as NLP, voice processing, computer vision, emotion analytics, and conversational systems, using different IBM Watson APIs. From this, the reader will learn what ML is, and what goes on in the background to make computers "do their magic," as well as where these concepts have been applied. Having achieved this, the readers will then be able to embark on their journey of learning, researching, and applying the concept in their respective fields. What you will learn Get well versed with the APIs provided by IBM Watson on IBM Cloud Learn ML, AI, cognitive computing, and neural network principles Implement smart applications in fields such as healthcare, entertainment, security, and more Understand unstructured content using cognitive metadata with the help of Natural Language Understanding Use Watson's APIs to create real-life applications to realize their capabilities Delve into various domains of cognitive computing, such as media analytics, embedded deep learning, computer vision, and more Who this book is forThis book is for beginners and novices; having some knowledge about artificial intelligence and deep learning is an advantage, but not a prerequisite to benefit from this book. We explain the concept of deep learning and artificial intelligence through the set of tools IBM Watson provides.
Quickly build and deploy machine learning models without managing infrastructure, and improve productivity using Amazon SageMaker's capabilities such as Amazon SageMaker Studio, Autopilot, Experiments, Debugger, and Model Monitor Key Features Build, train, and deploy machine learning models quickly using Amazon SageMaker Analyze, detect, and receive alerts relating to various business problems using machine learning algorithms and techniques Improve productivity by training and fine-tuning machine learning models in production Book DescriptionAmazon SageMaker enables you to quickly build, train, and deploy machine learning (ML) models at scale, without managing any infrastructure. It helps you focus on the ML problem at hand and deploy high-quality models by removing the heavy lifting typically involved in each step of the ML process. This book is a comprehensive guide for data scientists and ML developers who want to learn the ins and outs of Amazon SageMaker. You'll understand how to use various modules of SageMaker as a single toolset to solve the challenges faced in ML. As you progress, you'll cover features such as AutoML, built-in algorithms and frameworks, and the option for writing your own code and algorithms to build ML models. Later, the book will show you how to integrate Amazon SageMaker with popular deep learning libraries such as TensorFlow and PyTorch to increase the capabilities of existing models. You'll also learn to get the models to production faster with minimum effort and at a lower cost. Finally, you'll explore how to use Amazon SageMaker Debugger to analyze, detect, and highlight problems to understand the current model state and improve model accuracy. By the end of this Amazon book, you'll be able to use Amazon SageMaker on the full spectrum of ML workflows, from experimentation, training, and monitoring to scaling, deployment, and automation. What you will learn Create and automate end-to-end machine learning workflows on Amazon Web Services (AWS) Become well-versed with data annotation and preparation techniques Use AutoML features to build and train machine learning models with AutoPilot Create models using built-in algorithms and frameworks and your own code Train computer vision and NLP models using real-world examples Cover training techniques for scaling, model optimization, model debugging, and cost optimization Automate deployment tasks in a variety of configurations using SDK and several automation tools Who this book is forThis book is for software engineers, machine learning developers, data scientists, and AWS users who are new to using Amazon SageMaker and want to build high-quality machine learning models without worrying about infrastructure. Knowledge of AWS basics is required to grasp the concepts covered in this book more effectively. Some understanding of machine learning concepts and the Python programming language will also be beneficial.
People have described nature since the beginning of human history. They do it for various purposes, including to communicate about economic, social, governmental, meteorological, sustainability-related, strategic, military, and survival issues as well as artistic expression. As a part of the whole world of living beings, we use various types of senses, known and unknown, labeled and not identified, to both communicate and create. Describing Nature Through Visual Data is a collection of impactful research that discusses issues related to the visualization of scientific concepts, picturing processes, and products, as well as the role of computing in advancing visual literacy skills. Organized into four sections, the book contains descriptions, theories, and examples of visual and music-based solutions concerning the selected natural or technological events that are shaping present-day reality. The chapters pertain to selected scientific fields, digital art, computer graphics, and new media and confer the possible ways that visuals, visualization, simulation, and interactive knowledge presentation can help us to understand and share the content of scientific thought, research, artistic works, and practice. Featuring coverage on topics that include mathematical thinking, music theory, and visual communication, this reference is ideal for instructors, professionals, researchers, and students keen on comprehending and enhancing the role of knowledge visualization in computing, sciences, design, media communication, film, advertising, and marketing.
Learn to extract actionable insights from your big data in real time using a range of Microsoft Azure features Key Features Updated with the latest features and new additions to Microsoft Azure Master the fundamentals of cloud analytics using Azure Learn to use Azure Synapse Analytics (formerly known as Azure SQL Data Warehouse) to derive real-time customer insights Book DescriptionCloud Analytics with Microsoft Azure serves as a comprehensive guide for big data analysis and processing using a range of Microsoft Azure features. This book covers everything you need to build your own data warehouse and learn numerous techniques to gain useful insights by analyzing big data The book begins by introducing you to the power of data with big data analytics, the Internet of Things (IoT), machine learning, artificial intelligence, and DataOps. You will learn about cloud-scale analytics and the services Microsoft Azure offers to empower businesses to discover insights. You will also be introduced to the new features and functionalities added to the modern data warehouse. Finally, you will look at two real-world business use cases to demonstrate high-level solutions using Microsoft Azure. The aim of these use cases will be to illustrate how real-time data can be analyzed in Azure to derive meaningful insights and make business decisions. You will learn to build an end-to-end analytics pipeline on the cloud with machine learning and deep learning concepts. By the end of this book, you will be proficient in analyzing large amounts of data with Azure and using it effectively to benefit your organization. What you will learn Explore the concepts of modern data warehouses and data pipelines Discover unique design considerations while applying a cloud analytics solution Design an end-to-end analytics pipeline on the cloud Differentiate between structured, semi-structured, and unstructured data Choose a cloud-based service for your data analytics solutions Use Azure services to ingest, store, and analyze data of any scale Who this book is forThis book is designed to benefit software engineers, Azure developers, cloud consultants, and anyone who is keen to learn the process of deriving business insights from huge amounts of data using Azure. Though not necessary, a basic understanding of data analytics concepts such as data streaming, data types, the machine learning life cycle, and Docker containers will help you get the most out of the book.
Designed with beginners in mind, this workshop helps you make the most of Python libraries and the Jupyter Notebook's functionality to understand how data science can be applied to solve real-world data problems. Key Features Gain useful insights into data science and machine learning Explore the different functionalities and features of a Jupyter Notebook Discover how Python libraries are used with Jupyter for data analysis Book DescriptionFrom banking and manufacturing through to education and entertainment, using data science for business has revolutionized almost every sector in the modern world. It has an important role to play in everything from app development to network security. Taking an interactive approach to learning the fundamentals, this book is ideal for beginners. You'll learn all the best practices and techniques for applying data science in the context of real-world scenarios and examples. Starting with an introduction to data science and machine learning, you'll start by getting to grips with Jupyter functionality and features. You'll use Python libraries like sci-kit learn, pandas, Matplotlib, and Seaborn to perform data analysis and data preprocessing on real-world datasets from within your own Jupyter environment. Progressing through the chapters, you'll train classification models using sci-kit learn, and assess model performance using advanced validation techniques. Towards the end, you'll use Jupyter Notebooks to document your research, build stakeholder reports, and even analyze web performance data. By the end of The Applied Data Science Workshop, you'll be prepared to progress from being a beginner to taking your skills to the next level by confidently applying data science techniques and tools to real-world projects. What you will learn Understand the key opportunities and challenges in data science Use Jupyter for data science tasks such as data analysis and modeling Run exploratory data analysis within a Jupyter Notebook Visualize data with pairwise scatter plots and segmented distribution Assess model performance with advanced validation techniques Parse HTML responses and analyze HTTP requests Who this book is forIf you are an aspiring data scientist who wants to build a career in data science or a developer who wants to explore the applications of data science from scratch and analyze data in Jupyter using Python libraries, then this book is for you. Although a brief understanding of Python programming and machine learning is recommended to help you grasp the topics covered in the book more quickly, it is not mandatory.
Explore expert techniques such as advanced indexing and high availability to build scalable, reliable, and fault-tolerant database applications using PostgreSQL 13 Key Features Master advanced PostgreSQL 13 concepts with the help of real-world datasets and examples Leverage PostgreSQL's indexing features to fine-tune the performance of your queries Extend PostgreSQL's functionalities to suit your organization's needs with minimal effort Book DescriptionThanks to its reliability, robustness, and high performance, PostgreSQL has become one of the most advanced open source databases on the market. This updated fourth edition will help you understand PostgreSQL administration and how to build dynamic database solutions for enterprise apps with the latest release of PostgreSQL, including designing both physical and technical aspects of the system architecture with ease. Starting with an introduction to the new features in PostgreSQL 13, this book will guide you in building efficient and fault-tolerant PostgreSQL apps. You'll explore advanced PostgreSQL features, such as logical replication, database clusters, performance tuning, advanced indexing, monitoring, and user management, to manage and maintain your database. You'll then work with the PostgreSQL optimizer, configure PostgreSQL for high speed, and move from Oracle to PostgreSQL. The book also covers transactions, locking, and indexes, and shows you how to improve performance with query optimization. You'll also focus on how to manage network security and work with backups and replication while exploring useful PostgreSQL extensions that optimize the performance of large databases. By the end of this PostgreSQL book, you'll be able to get the most out of your database by executing advanced administrative tasks. What you will learn Get well versed with advanced SQL functions in PostgreSQL 13 Get to grips with administrative tasks such as log file management and monitoring Work with stored procedures and manage backup and recovery Employ replication and failover techniques to reduce data loss Perform database migration from Oracle to PostgreSQL with ease Replicate PostgreSQL database systems to create backups and scale your database Manage and improve server security to protect your data Troubleshoot your PostgreSQL instance to find solutions to common and not-so-common problems Who this book is forThis database administration book is for PostgreSQL developers and database administrators and professionals who want to implement advanced functionalities and master complex administrative tasks with PostgreSQL 13. Prior experience in PostgreSQL and familiarity with the basics of database administration will assist with understanding key concepts covered in the book.
Get to grips with pandas by working with real datasets and master data discovery, data manipulation, data preparation, and handling data for analytical tasks Key Features Perform efficient data analysis and manipulation tasks using pandas 1.x Apply pandas to different real-world domains with the help of step-by-step examples Make the most of pandas as an effective data exploration tool Book DescriptionExtracting valuable business insights is no longer a 'nice-to-have', but an essential skill for anyone who handles data in their enterprise. Hands-On Data Analysis with Pandas is here to help beginners and those who are migrating their skills into data science get up to speed in no time. This book will show you how to analyze your data, get started with machine learning, and work effectively with the Python libraries often used for data science, such as pandas, NumPy, matplotlib, seaborn, and scikit-learn. Using real-world datasets, you will learn how to use the pandas library to perform data wrangling to reshape, clean, and aggregate your data. Then, you will learn how to conduct exploratory data analysis by calculating summary statistics and visualizing the data to find patterns. In the concluding chapters, you will explore some applications of anomaly detection, regression, clustering, and classification using scikit-learn to make predictions based on past data. This updated edition will equip you with the skills you need to use pandas 1.x to efficiently perform various data manipulation tasks, reliably reproduce analyses, and visualize your data for effective decision making - valuable knowledge that can be applied across multiple domains. What you will learn Understand how data analysts and scientists gather and analyze data Perform data analysis and data wrangling using Python Combine, group, and aggregate data from multiple sources Create data visualizations with pandas, matplotlib, and seaborn Apply machine learning algorithms to identify patterns and make predictions Use Python data science libraries to analyze real-world datasets Solve common data representation and analysis problems using pandas Build Python scripts, modules, and packages for reusable analysis code Who this book is forThis book is for data science beginners, data analysts, and Python developers who want to explore each stage of data analysis and scientific computing using a wide range of datasets. Data scientists looking to implement pandas in their machine learning workflow will also find plenty of valuable know-how as they progress. You'll find it easier to follow along with this book if you have a working knowledge of the Python programming language, but a Python crash-course tutorial is provided in the code bundle for anyone who needs a refresher.
Quickly build and deploy massive data pipelines and improve productivity using Azure Databricks Key Features Get to grips with the distributed training and deployment of machine learning and deep learning models Learn how ETLs are integrated with Azure Data Factory and Delta Lake Explore deep learning and machine learning models in a distributed computing infrastructure Book DescriptionMicrosoft Azure Databricks helps you to harness the power of distributed computing and apply it to create robust data pipelines, along with training and deploying machine learning and deep learning models. Databricks' advanced features enable developers to process, transform, and explore data. Distributed Data Systems with Azure Databricks will help you to put your knowledge of Databricks to work to create big data pipelines. The book provides a hands-on approach to implementing Azure Databricks and its associated methodologies that will make you productive in no time. Complete with detailed explanations of essential concepts, practical examples, and self-assessment questions, you'll begin with a quick introduction to Databricks core functionalities, before performing distributed model training and inference using TensorFlow and Spark MLlib. As you advance, you'll explore MLflow Model Serving on Azure Databricks and implement distributed training pipelines using HorovodRunner in Databricks. Finally, you'll discover how to transform, use, and obtain insights from massive amounts of data to train predictive models and create entire fully working data pipelines. By the end of this MS Azure book, you'll have gained a solid understanding of how to work with Databricks to create and manage an entire big data pipeline. What you will learn Create ETLs for big data in Azure Databricks Train, manage, and deploy machine learning and deep learning models Integrate Databricks with Azure Data Factory for extract, transform, load (ETL) pipeline creation Discover how to use Horovod for distributed deep learning Find out how to use Delta Engine to query and process data from Delta Lake Understand how to use Data Factory in combination with Databricks Use Structured Streaming in a production-like environment Who this book is forThis book is for software engineers, machine learning engineers, data scientists, and data engineers who are new to Azure Databricks and want to build high-quality data pipelines without worrying about infrastructure. Knowledge of Azure Databricks basics is required to learn the concepts covered in this book more effectively. A basic understanding of machine learning concepts and beginner-level Python programming knowledge is also recommended.
Work through practical recipes to learn how to solve complex machine learning and deep learning problems using Python Key Features Get up and running with artificial intelligence in no time using hands-on problem-solving recipes Explore popular Python libraries and tools to build AI solutions for images, text, sounds, and images Implement NLP, reinforcement learning, deep learning, GANs, Monte-Carlo tree search, and much more Book DescriptionArtificial intelligence (AI) plays an integral role in automating problem-solving. This involves predicting and classifying data and training agents to execute tasks successfully. This book will teach you how to solve complex problems with the help of independent and insightful recipes ranging from the essentials to advanced methods that have just come out of research. Artificial Intelligence with Python Cookbook starts by showing you how to set up your Python environment and taking you through the fundamentals of data exploration. Moving ahead, you'll be able to implement heuristic search techniques and genetic algorithms. In addition to this, you'll apply probabilistic models, constraint optimization, and reinforcement learning. As you advance through the book, you'll build deep learning models for text, images, video, and audio, and then delve into algorithmic bias, style transfer, music generation, and AI use cases in the healthcare and insurance industries. Throughout the book, you'll learn about a variety of tools for problem-solving and gain the knowledge needed to effectively approach complex problems. By the end of this book on AI, you will have the skills you need to write AI and machine learning algorithms, test them, and deploy them for production. What you will learn Implement data preprocessing steps and optimize model hyperparameters Delve into representational learning with adversarial autoencoders Use active learning, recommenders, knowledge embedding, and SAT solvers Get to grips with probabilistic modeling with TensorFlow probability Run object detection, text-to-speech conversion, and text and music generation Apply swarm algorithms, multi-agent systems, and graph networks Go from proof of concept to production by deploying models as microservices Understand how to use modern AI in practice Who this book is forThis AI machine learning book is for Python developers, data scientists, machine learning engineers, and deep learning practitioners who want to learn how to build artificial intelligence solutions with easy-to-follow recipes. You'll also find this book useful if you're looking for state-of-the-art solutions to perform different machine learning tasks in various use cases. Basic working knowledge of the Python programming language and machine learning concepts will help you to work with code effectively in this book.
Understand data analysis pipelines using machine learning algorithms and techniques with this practical guide Key Features Prepare and clean your data to use it for exploratory analysis, data manipulation, and data wrangling Discover supervised, unsupervised, probabilistic, and Bayesian machine learning methods Get to grips with graph processing and sentiment analysis Book DescriptionData analysis enables you to generate value from small and big data by discovering new patterns and trends, and Python is one of the most popular tools for analyzing a wide variety of data. With this book, you'll get up and running using Python for data analysis by exploring the different phases and methodologies used in data analysis and learning how to use modern libraries from the Python ecosystem to create efficient data pipelines. Starting with the essential statistical and data analysis fundamentals using Python, you'll perform complex data analysis and modeling, data manipulation, data cleaning, and data visualization using easy-to-follow examples. You'll then understand how to conduct time series analysis and signal processing using ARMA models. As you advance, you'll get to grips with smart processing and data analytics using machine learning algorithms such as regression, classification, Principal Component Analysis (PCA), and clustering. In the concluding chapters, you'll work on real-world examples to analyze textual and image data using natural language processing (NLP) and image analytics techniques, respectively. Finally, the book will demonstrate parallel computing using Dask. By the end of this data analysis book, you'll be equipped with the skills you need to prepare data for analysis and create meaningful data visualizations for forecasting values from data. What you will learn Explore data science and its various process models Perform data manipulation using NumPy and pandas for aggregating, cleaning, and handling missing values Create interactive visualizations using Matplotlib, Seaborn, and Bokeh Retrieve, process, and store data in a wide range of formats Understand data preprocessing and feature engineering using pandas and scikit-learn Perform time series analysis and signal processing using sunspot cycle data Analyze textual data and image data to perform advanced analysis Get up to speed with parallel computing using Dask Who this book is forThis book is for data analysts, business analysts, statisticians, and data scientists looking to learn how to use Python for data analysis. Students and academic faculties will also find this book useful for learning and teaching Python data analysis using a hands-on approach. A basic understanding of math and working knowledge of the Python programming language will help you get started with this book.
Informationsgesellschaft, Information als Wettbewerbsfaktor, Informationsflut: Diese Stichworte verdeutlichen die unternehmerische und gesellschaftliche Bedeutung von Informationen. Doch nicht nur Information allein, sondern auch die Systeme, die Informationen verarbeiten, speichern und ubertragen sowie die Technologien, auf denen sie beruhen, verdienen Aufmerksamkeit. Informationsmanagement hat die Aufgabe, den im Hinblick auf das Unternehmensziel bestmoeglichen Einsatz der Ressource Information zu gewahrleisten. Es zahlt zu den wesentlichen Bestandteilen heutiger Unternehmensfuhrung. Das Lehrbuch vermittelt in 13 Einheiten die Grundlagen des Informationsmanagements. Dabei werden neben den Managementaufgaben der Informationswirtschaft, der Systeme und der Technologien auch ausgewahlte Fuhrungsaufgaben des Informationsmanagementsbehandelt. Jede Lehreinheit beginnt mit einem UEberblick uber die behandelten Themen und schliesst mit einer Zusammenfassung sowie Aufgaben zur Wiederholung ab. So richtet sich dieses Buch insbesondere an Bachelorstudenten in den Fachern Wirtschaftsinformatik, BWL und Informatik. |
You may like...
Using Technology to Enhance Special…
Jeffrey P. Bakken, Festus E. Obiakor
Hardcover
R3,378
Discovery Miles 33 780
Future Learning in Primary Schools - A…
Ching Sing CHAI, Cher Ping Lim, …
Hardcover
The Educational Intelligent Economy…
Tavis D. Jules, Florin D. Salajan
Hardcover
R2,858
Discovery Miles 28 580
Leveraging Technology to Improve School…
Stephanie P. Huffman, Stacey Loyless, …
Hardcover
R5,372
Discovery Miles 53 720
Technology-Driven E-Learning Pedagogy…
Pooja Chaturvedi Sharma, Rohit Bansal, …
Hardcover
R5,372
Discovery Miles 53 720
Facilitating Learning in Language…
Rajest S. Suman, Salvatore Moccia, …
Hardcover
R6,677
Discovery Miles 66 770
Institutional Transformation Through…
Mark Stansfield, Thomas Connolly
Hardcover
R4,942
Discovery Miles 49 420
|