![]() |
Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
||
|
Books > Computing & IT > Applications of computing > Databases > Data capture & analysis
Learn how to architect, implement, and administer a complex Splunk Enterprise environment and extract valuable insights from business data. Key Features Understand the various components of Splunk and how they work together to provide a powerful Big Data analytics solution. Collect and index data from a wide variety of common machine data sources Design searches, reports, and dashboard visualizations to provide business data insights Book DescriptionSplunk is a leading platform and solution for collecting, searching, and extracting value from ever increasing amounts of big data - and big data is eating the world! This book covers all the crucial Splunk topics and gives you the information and examples to get the immediate job done. You will find enough insights to support further research and use Splunk to suit any business environment or situation. Splunk 7.x Quick Start Guide gives you a thorough understanding of how Splunk works. You will learn about all the critical tasks for architecting, implementing, administering, and utilizing Splunk Enterprise to collect, store, retrieve, format, analyze, and visualize machine data. You will find step-by-step examples based on real-world experience and practical use cases that are applicable to all Splunk environments. There is a careful balance between adequate coverage of all the critical topics with short but relevant deep-dives into the configuration options and steps to carry out the day-to-day tasks that matter. By the end of the book, you will be a confident and proficient Splunk architect and administrator. What you will learn Design and implement a complex Splunk Enterprise solution Configure your Splunk environment to get machine data in and indexed Build searches to get and format data for analysis and visualization Build reports, dashboards, and alerts to deliver critical insights Create knowledge objects to enhance the value of your data Install Splunk apps to provide focused views into key technologies Monitor, troubleshoot, and manage your Splunk environment Who this book is forThis book is intended for experienced IT personnel who are just getting started working with Splunk and want to quickly become proficient with its usage. Data analysts who need to leverage Splunk to extract critical business insights from application logs and other machine data sources will also benefit from this book.
Extend PostgreSQL using PostgreSQL server programming to create, test, debug, and optimize a range of user-defined functions in your favorite programming language Key Features Learn the concepts of PostgreSQL 11 with lots of real-world datasets and examples Learn queries, data replication, and database performance Extend the functionalities of your PostgreSQL instance to suit your organizational needs Book DescriptionPostgreSQL is a rock-solid, scalable, and safe enterprise-level relational database. With a broad range of features and stability, it is ever increasing in popularity.This book shows you how to take advantage of PostgreSQL 11 features for server-side programming. Server-side programming enables strong data encapsulation and coherence. The book begins with the importance of server-side programming and explains the risks of leaving all the checks outside the database. To build your capabilities further, you will learn how to write stored procedures, both functions and the new PostgreSQL 11 procedures, and create triggers to perform encapsulation and maintain data consistency. You will also learn how to produce extensions, the easiest way to package your programs for easy and solid deployment on different PostgreSQL installations. What you will learn Explore data encapsulation Write stored procedures in different languages Interact with transactions from within a function Get to grips with triggers and rules Create and manage custom data types Create extensions to package code and data Implement background workers and Inter-Process Communication (IPC) How to deal with foreign languages, in particular Java and Perl Who this book is forThis book is for database administrators, data engineers, and database engineers who want to implement advanced functionalities and master complex administrative tasks with PostgreSQL 11.
Get up to speed with the new features added to Microsoft SQL Server 2019 Analysis Services and create models to support your business Key Features Explore tips and tricks to design, develop, and optimize end-to-end data analytics solutions using Microsoft's technologies Learn tabular modeling and multi-dimensional cube design development using real-world examples Implement Analysis Services to help you make productive business decisions Book DescriptionSQL Server Analysis Services (SSAS) continues to be a leading enterprise-scale toolset, enabling customers to deliver data and analytics across large datasets with great performance. This book will help you understand MS SQL Server 2019's new features and improvements, especially when it comes to SSAS. First, you'll cover a quick overview of SQL Server 2019, learn how to choose the right analytical model to use, and understand their key differences. You'll then explore how to create a multi-dimensional model with SSAS and expand on that model with MDX. Next, you'll create and deploy a tabular model using Microsoft Visual Studio and Management Studio. You'll learn when and how to use both tabular and multi-dimensional model types, how to deploy and configure your servers to support them, and design principles that are relevant to each model. The book comes packed with tips and tricks to build measures, optimize your design, and interact with models using Excel and Power BI. All this will help you visualize data to gain useful insights and make better decisions. Finally, you'll discover practices and tools for securing and maintaining your models once they are deployed. By the end of this MS SQL Server book, you'll be able to choose the right model and build and deploy it to support the analytical needs of your business. What you will learn Determine the best analytical model using SSAS Cover the core aspects involved in MDX, including writing your first query Implement calculated tables and calculation groups (new in version 2019) in DAX Create and deploy tabular and multi-dimensional models on SQL 2019 Connect and create data visualizations using Excel and Power BI Implement row-level and other data security methods with tabular and multi-dimensional models Explore essential concepts and techniques to scale, manage, and optimize your SSAS solutions Who this book is forThis Microsoft SQL Server book is for BI professionals and data analysts who are looking for a practical guide to creating and maintaining tabular and multi-dimensional models using SQL Server 2019 Analysis Services. A basic working knowledge of BI solutions such as Power BI and database querying is required.
Kickstart your NLP journey by exploring BERT and its variants such as ALBERT, RoBERTa, DistilBERT, VideoBERT, and more with Hugging Face's transformers library Key Features Explore the encoder and decoder of the transformer model Become well-versed with BERT along with ALBERT, RoBERTa, and DistilBERT Discover how to pre-train and fine-tune BERT models for several NLP tasks Book DescriptionBERT (bidirectional encoder representations from transformer) has revolutionized the world of natural language processing (NLP) with promising results. This book is an introductory guide that will help you get to grips with Google's BERT architecture. With a detailed explanation of the transformer architecture, this book will help you understand how the transformer's encoder and decoder work. You'll explore the BERT architecture by learning how the BERT model is pre-trained and how to use pre-trained BERT for downstream tasks by fine-tuning it for NLP tasks such as sentiment analysis and text summarization with the Hugging Face transformers library. As you advance, you'll learn about different variants of BERT such as ALBERT, RoBERTa, and ELECTRA, and look at SpanBERT, which is used for NLP tasks like question answering. You'll also cover simpler and faster BERT variants based on knowledge distillation such as DistilBERT and TinyBERT. The book takes you through MBERT, XLM, and XLM-R in detail and then introduces you to sentence-BERT, which is used for obtaining sentence representation. Finally, you'll discover domain-specific BERT models such as BioBERT and ClinicalBERT, and discover an interesting variant called VideoBERT. By the end of this BERT book, you'll be well-versed with using BERT and its variants for performing practical NLP tasks. What you will learn Understand the transformer model from the ground up Find out how BERT works and pre-train it using masked language model (MLM) and next sentence prediction (NSP) tasks Get hands-on with BERT by learning to generate contextual word and sentence embeddings Fine-tune BERT for downstream tasks Get to grips with ALBERT, RoBERTa, ELECTRA, and SpanBERT models Get the hang of the BERT models based on knowledge distillation Understand cross-lingual models such as XLM and XLM-R Explore Sentence-BERT, VideoBERT, and BART Who this book is forThis book is for NLP professionals and data scientists looking to simplify NLP tasks to enable efficient language understanding using BERT. A basic understanding of NLP concepts and deep learning is required to get the best out of this book.
Get to grips with automated machine learning and adopt a hands-on approach to AutoML implementation and associated methodologies Key Features Get up to speed with AutoML using OSS, Azure, AWS, GCP, or any platform of your choice Eliminate mundane tasks in data engineering and reduce human errors in machine learning models Find out how you can make machine learning accessible for all users to promote decentralized processes Book DescriptionEvery machine learning engineer deals with systems that have hyperparameters, and the most basic task in automated machine learning (AutoML) is to automatically set these hyperparameters to optimize performance. The latest deep neural networks have a wide range of hyperparameters for their architecture, regularization, and optimization, which can be customized effectively to save time and effort. This book reviews the underlying techniques of automated feature engineering, model and hyperparameter tuning, gradient-based approaches, and much more. You'll discover different ways of implementing these techniques in open source tools and then learn to use enterprise tools for implementing AutoML in three major cloud service providers: Microsoft Azure, Amazon Web Services (AWS), and Google Cloud Platform. As you progress, you'll explore the features of cloud AutoML platforms by building machine learning models using AutoML. The book will also show you how to develop accurate models by automating time-consuming and repetitive tasks in the machine learning development lifecycle. By the end of this machine learning book, you'll be able to build and deploy AutoML models that are not only accurate, but also increase productivity, allow interoperability, and minimize feature engineering tasks. What you will learn Explore AutoML fundamentals, underlying methods, and techniques Assess AutoML aspects such as algorithm selection, auto featurization, and hyperparameter tuning in an applied scenario Find out the difference between cloud and operations support systems (OSS) Implement AutoML in enterprise cloud to deploy ML models and pipelines Build explainable AutoML pipelines with transparency Understand automated feature engineering and time series forecasting Automate data science modeling tasks to implement ML solutions easily and focus on more complex problems Who this book is forCitizen data scientists, machine learning developers, artificial intelligence enthusiasts, or anyone looking to automatically build machine learning models using the features offered by open source tools, Microsoft Azure Machine Learning, AWS, and Google Cloud Platform will find this book useful. Beginner-level knowledge of building ML models is required to get the best out of this book. Prior experience in using Enterprise cloud is beneficial.
Dieses Buch bietet einen historisch orientierten Einstieg in die elementare Zahlentheorie. Es stellt eine solide Basis fur vielfaltige Ausbaumoeglichkeiten dar. Besondere Zielsetzungen sind: Elementaritat und Anschaulichkeit, die Berucksichtigung der historischen Entwicklung, Motivation der Begriffe und Verfahren anhand konkreter, aussagekraftiger Beispiele unter Einbezug moderner Werkzeuge (Computeralgebra Systeme, Internet). Als Zusatzmedien werden Computer- und Internet-spezifische Interaktions- und Visualisierungsmoeglichkeiten (kostenlos) zur Verfugung gestellt. Das Werk wendet sich an Studierende (Bachelor/Lehramt), Lehrer(innen) sowie alle an Elementarmathematik interessierten Leser.
Discover how to use Neo4j to identify relationships within complex and large graph datasets using graph modeling, graph algorithms, and machine learning Key Features Get up and running with graph analytics with the help of real-world examples Explore various use cases such as fraud detection, graph-based search, and recommendation systems Get to grips with the Graph Data Science library with the help of examples, and use Neo4j in the cloud for effective application scaling Book DescriptionNeo4j is a graph database that includes plugins to run complex graph algorithms. The book starts with an introduction to the basics of graph analytics, the Cypher query language, and graph architecture components, and helps you to understand why enterprises have started to adopt graph analytics within their organizations. You'll find out how to implement Neo4j algorithms and techniques and explore various graph analytics methods to reveal complex relationships in your data. You'll be able to implement graph analytics catering to different domains such as fraud detection, graph-based search, recommendation systems, social networking, and data management. You'll also learn how to store data in graph databases and extract valuable insights from it. As you become well-versed with the techniques, you'll discover graph machine learning in order to address simple to complex challenges using Neo4j. You will also understand how to use graph data in a machine learning model in order to make predictions based on your data. Finally, you'll get to grips with structuring a web application for production using Neo4j. By the end of this book, you'll not only be able to harness the power of graphs to handle a broad range of problem areas, but you'll also have learned how to use Neo4j efficiently to identify complex relationships in your data. What you will learn Become well-versed with Neo4j graph database building blocks, nodes, and relationships Discover how to create, update, and delete nodes and relationships using Cypher querying Use graphs to improve web search and recommendations Understand graph algorithms such as pathfinding, spatial search, centrality, and community detection Find out different steps to integrate graphs in a normal machine learning pipeline Formulate a link prediction problem in the context of machine learning Implement graph embedding algorithms such as DeepWalk, and use them in Neo4j graphs Who this book is forThis book is for data analysts, business analysts, graph analysts, and database developers looking to store and process graph data to reveal key data insights. This book will also appeal to data scientists who want to build intelligent graph applications catering to different domains. Some experience with Neo4j is required.
A practical guide to getting the most out of Excel, using it for data preparation, applying machine learning models (including cloud services) and understanding the outcome of the data analysis. Key Features Use Microsoft's product Excel to build advanced forecasting models using varied examples Cover range of machine learning tasks such as data mining, data analytics, smart visualization, and more Derive data-driven techniques using Excel plugins and APIs without much code required Book DescriptionWe have made huge progress in teaching computers to perform difficult tasks, especially those that are repetitive and time-consuming for humans. Excel users, of all levels, can feel left behind by this innovation wave. The truth is that a large amount of the work needed to develop and use a machine learning model can be done in Excel. The book starts by giving a general introduction to machine learning, making every concept clear and understandable. Then, it shows every step of a machine learning project, from data collection, reading from different data sources, developing models, and visualizing the results using Excel features and offerings. In every chapter, there are several examples and hands-on exercises that will show the reader how to combine Excel functions, add-ins, and connections to databases and to cloud services to reach the desired goal: building a full data analysis flow. Different machine learning models are shown, tailored to the type of data to be analyzed. At the end of the book, the reader is presented with some advanced use cases using Automated Machine Learning, and artificial neural network, which simplifies the analysis task and represents the future of machine learning. What you will learn Use Excel to preview and cleanse datasets Understand correlations between variables and optimize the input to machine learning models Use and evaluate different machine learning models from Excel Understand the use of different visualizations Learn the basic concepts and calculations to understand how artificial neural networks work Learn how to connect Excel to the Microsoft Azure cloud Get beyond proof of concepts and build fully functional data analysis flows Who this book is forThis book is for data analysis, machine learning enthusiasts, project managers, and someone who doesn't want to code much for performing core tasks of machine learning. Each example will help you perform end-to-end smart analytics. Working knowledge of Excel is required.
Web Scraping techniques are getting more popular, since data is as valuable as oil in 21st century. Through this book get some key knowledge about using XPath, regEX; web scraping libraries for R like rvest and RSelenium technologies. Key Features Techniques, tools and frameworks for web scraping with R Scrape data effortlessly from a variety of websites Learn how to selectively choose the data to scrape, and build your dataset Book DescriptionWeb scraping is a technique to extract data from websites. It simulates the behavior of a website user to turn the website itself into a web service to retrieve or introduce new data. This book gives you all you need to get started with scraping web pages using R programming. You will learn about the rules of RegEx and Xpath, key components for scraping website data. We will show you web scraping techniques, methodologies, and frameworks. With this book's guidance, you will become comfortable with the tools to write and test RegEx and XPath rules. We will focus on examples of dynamic websites for scraping data and how to implement the techniques learned. You will learn how to collect URLs and then create XPath rules for your first web scraping script using rvest library. From the data you collect, you will be able to calculate the statistics and create R plots to visualize them. Finally, you will discover how to use Selenium drivers with R for more sophisticated scraping. You will create AWS instances and use R to connect a PostgreSQL database hosted on AWS. By the end of the book, you will be sufficiently confident to create end-to-end web scraping systems using R. What you will learn Write and create regEX rules Write XPath rules to query your data Learn how web scraping methods work Use rvest to crawl web pages Store data retrieved from the web Learn the key uses of Rselenium to scrape data Who this book is forThis book is for R programmers who want to get started quickly with web scraping, as well as data analysts who want to learn scraping using R. Basic knowledge of R is all you need to get started with this book.
Understand the fundamentals of algorithmic trading to apply algorithms to real market data and analyze the results of real-world trading strategies Key Features Understand the power of algorithmic trading in financial markets with real-world examples Get up and running with the algorithms used to carry out algorithmic trading Learn to build your own algorithmic trading robots which require no human intervention Book DescriptionIt's now harder than ever to get a significant edge over competitors in terms of speed and efficiency when it comes to algorithmic trading. Relying on sophisticated trading signals, predictive models and strategies can make all the difference. This book will guide you through these aspects, giving you insights into how modern electronic trading markets and participants operate. You'll start with an introduction to algorithmic trading, along with setting up the environment required to perform the tasks in the book. You'll explore the key components of an algorithmic trading business and aspects you'll need to take into account before starting an automated trading project. Next, you'll focus on designing, building and operating the components required for developing a practical and profitable algorithmic trading business. Later, you'll learn how quantitative trading signals and strategies are developed, and also implement and analyze sophisticated trading strategies such as volatility strategies, economic release strategies, and statistical arbitrage. Finally, you'll create a trading bot from scratch using the algorithms built in the previous sections. By the end of this book, you'll be well-versed with electronic trading markets and have learned to implement, evaluate and safely operate algorithmic trading strategies in live markets. What you will learn Understand the components of modern algorithmic trading systems and strategies Apply machine learning in algorithmic trading signals and strategies using Python Build, visualize and analyze trading strategies based on mean reversion, trend, economic releases and more Quantify and build a risk management system for Python trading strategies Build a backtester to run simulated trading strategies for improving the performance of your trading bot Deploy and incorporate trading strategies in the live market to maintain and improve profitability Who this book is forThis book is for software engineers, financial traders, data analysts, and entrepreneurs. Anyone who wants to get started with algorithmic trading and understand how it works; and learn the components of a trading system, protocols and algorithms required for black box and gray box trading, and techniques for building a completely automated and profitable trading business will also find this book useful.
Simplify your ETL processes with these hands-on data hygiene tips, tricks, and best practices. Key Features Focus on the basics of data wrangling Study various ways to extract the most out of your data in less time Boost your learning curve with bonus topics like random data generation and data integrity checks Book DescriptionFor data to be useful and meaningful, it must be curated and refined. Data Wrangling with Python teaches you the core ideas behind these processes and equips you with knowledge of the most popular tools and techniques in the domain. The book starts with the absolute basics of Python, focusing mainly on data structures. It then delves into the fundamental tools of data wrangling like NumPy and Pandas libraries. You'll explore useful insights into why you should stay away from traditional ways of data cleaning, as done in other languages, and take advantage of the specialized pre-built routines in Python. This combination of Python tips and tricks will also demonstrate how to use the same Python backend and extract/transform data from an array of sources including the Internet, large database vaults, and Excel financial tables. To help you prepare for more challenging scenarios, you'll cover how to handle missing or wrong data, and reformat it based on the requirements from the downstream analytics tool. The book will further help you grasp concepts through real-world examples and datasets. By the end of this book, you will be confident in using a diverse array of sources to extract, clean, transform, and format your data efficiently. What you will learn Use and manipulate complex and simple data structures Harness the full potential of DataFrames and numpy.array at run time Perform web scraping with BeautifulSoup4 and html5lib Execute advanced string search and manipulation with RegEX Handle outliers and perform data imputation with Pandas Use descriptive statistics and plotting techniques Practice data wrangling and modeling using data generation techniques Who this book is forData Wrangling with Python is designed for developers, data analysts, and business analysts who are keen to pursue a career as a full-fledged data scientist or analytics expert. Although, this book is for beginners, prior working knowledge of Python is necessary to easily grasp the concepts covered here. It will also help to have rudimentary knowledge of relational database and SQL.
Build a solid foundation in algorithmic trading by developing, testing and executing powerful trading strategies with real market data using Python Key Features Build a strong foundation in algorithmic trading by becoming well-versed with the basics of financial markets Demystify jargon related to understanding and placing multiple types of trading orders Devise trading strategies and increase your odds of making a profit without human intervention Book DescriptionIf you want to find out how you can build a solid foundation in algorithmic trading using Python, this cookbook is here to help. Starting by setting up the Python environment for trading and connectivity with brokers, you'll then learn the important aspects of financial markets. As you progress, you'll learn to fetch financial instruments, query and calculate various types of candles and historical data, and finally, compute and plot technical indicators. Next, you'll learn how to place various types of orders, such as regular, bracket, and cover orders, and understand their state transitions. Later chapters will cover backtesting, paper trading, and finally real trading for the algorithmic strategies that you've created. You'll even understand how to automate trading and find the right strategy for making effective decisions that would otherwise be impossible for human traders. By the end of this book, you'll be able to use Python libraries to conduct key tasks in the algorithmic trading ecosystem. Note: For demonstration, we're using Zerodha, an Indian Stock Market broker. If you're not an Indian resident, you won't be able to use Zerodha and therefore will not be able to test the examples directly. However, you can take inspiration from the book and apply the concepts across your preferred stock market broker of choice. What you will learn Use Python to set up connectivity with brokers Handle and manipulate time series data using Python Fetch a list of exchanges, segments, financial instruments, and historical data to interact with the real market Understand, fetch, and calculate various types of candles and use them to compute and plot diverse types of technical indicators Develop and improve the performance of algorithmic trading strategies Perform backtesting and paper trading on algorithmic trading strategies Implement real trading in the live hours of stock markets Who this book is forIf you are a financial analyst, financial trader, data analyst, algorithmic trader, trading enthusiast or anyone who wants to learn algorithmic trading with Python and important techniques to address challenges faced in the finance domain, this book is for you. Basic working knowledge of the Python programming language is expected. Although fundamental knowledge of trade-related terminologies will be helpful, it is not mandatory.
Leverage the power of Tableau to get actionable business insights and make better business decisions Key Features Explore all the new features of Tableau 2018.x Begin your Tableau journey by understanding its core functionalities Apply advanced techniques such as data blending and parameters to make the best of Tableau Book DescriptionTableau is one of the leading business intelligence tools used worldwide, in organizations of every scale. In its latest release, Tableau 2018 promises richer and more useful features related to visual analytics, reporting, dashboarding, and a host of other data visualization aspects. Getting Started with Tableau 2018.x will get you up and running with these features. The book starts with all the new functionalities of the different Tableau 2018 versions, along with concrete examples of how to use them. However, if you're new to Tableau, don't worry! The rest of the book will guide you through each major aspect of Tableau with examples. You'll learn how to connect to data, build a data source, visualize your data, build a dashboard, and share it online. In the final chapters, you'll also learn advanced techniques such as creating a cross-database join, data blending, and more. By the end of the book, you will have a firm understanding of how to effectively use Tableau to create quick, cost-effective, and business-efficient business intelligence solutions. What you will learn Discover new functionalities such as density, extensions, and transparency introduced in Tableau 2018.x Connect tables and make transformations such as pivoting the field and splitting columns Build an efficient data source for your analysis Design insightful data visualization using different mark types and properties Build powerful dashboards and stories Share your work and interact with Tableau Server Use Tableau to explore your data and find new insights Explore Tableau's advanced features and gear up for upcoming challenges Who this book is forIf you're a beginner or an existing BI professional looking to get the most out of Tableau 2018's latest features, this book is for you. You'll also find this book useful if you're an aspiring analyst who wants to learn the capabilities of Tableau to answer business-specific questions. No prior knowledge or experience with Tableau or any other Business Intelligence tool is necessary.
Reinforce your understanding of data science and data analysis from a statistical perspective to extract meaningful insights from your data using Python programming Key Features Work your way through the entire data analysis pipeline with statistics concerns in mind to make reasonable decisions Understand how various data science algorithms function Build a solid foundation in statistics for data science and machine learning using Python-based examples Book DescriptionStatistics remain the backbone of modern analysis tasks, helping you to interpret the results produced by data science pipelines. This book is a detailed guide covering the math and various statistical methods required for undertaking data science tasks. The book starts by showing you how to preprocess data and inspect distributions and correlations from a statistical perspective. You'll then get to grips with the fundamentals of statistical analysis and apply its concepts to real-world datasets. As you advance, you'll find out how statistical concepts emerge from different stages of data science pipelines, understand the summary of datasets in the language of statistics, and use it to build a solid foundation for robust data products such as explanatory models and predictive models. Once you've uncovered the working mechanism of data science algorithms, you'll cover essential concepts for efficient data collection, cleaning, mining, visualization, and analysis. Finally, you'll implement statistical methods in key machine learning tasks such as classification, regression, tree-based methods, and ensemble learning. By the end of this Essential Statistics for Non-STEM Data Analysts book, you'll have learned how to build and present a self-contained, statistics-backed data product to meet your business goals. What you will learn Find out how to grab and load data into an analysis environment Perform descriptive analysis to extract meaningful summaries from data Discover probability, parameter estimation, hypothesis tests, and experiment design best practices Get to grips with resampling and bootstrapping in Python Delve into statistical tests with variance analysis, time series analysis, and A/B test examples Understand the statistics behind popular machine learning algorithms Answer questions on statistics for data scientist interviews Who this book is forThis book is an entry-level guide for data science enthusiasts, data analysts, and anyone starting out in the field of data science and looking to learn the essential statistical concepts with the help of simple explanations and examples. If you're a developer or student with a non-mathematical background, you'll find this book useful. Working knowledge of the Python programming language is required.
Enter the exciting world of Julia, a high-performance language for technical computing Key Features Leverage Julia's high speed and efficiency for your applications Work with Julia in a multi-core, distributed, and networked environment Apply Julia to tackle problems concurrently and in a distributed environment Book DescriptionThe release of Julia 1.0 is now ready to change the technical world by combining the high productivity and ease of use of Python and R with the lightning-fast speed of C++. Julia 1.0 programming gives you a head start in tackling your numerical and data problems. You will begin by learning how to set up a running Julia platform, before exploring its various built-in types. With the help of practical examples, this book walks you through two important collection types: arrays and matrices. In addition to this, you will be taken through how type conversions and promotions work. In the course of the book, you will be introduced to the homo-iconicity and metaprogramming concepts in Julia. You will understand how Julia provides different ways to interact with an operating system, as well as other languages, and then you'll discover what macros are. Once you have grasped the basics, you'll study what makes Julia suitable for numerical and scientific computing, and learn about the features provided by Julia. By the end of this book, you will also have learned how to run external programs. This book covers all you need to know about Julia in order to leverage its high speed and efficiency for your applications. What you will learn Set up your Julia environment to achieve high productivity Create your own types to extend the built-in type system Visualize your data in Julia with plotting packages Explore the use of built-in macros for testing and debugging, among other uses Apply Julia to tackle problems concurrently Integrate Julia with other languages such as C, Python, and MATLAB Who this book is forJulia 1.0 Programming is for you if you are a statistician or data scientist who wants a crash course in the Julia programming language while building big data applications. A basic knowledge of mathematics is needed to understand the various methods that are used or created during the course of the book to exploit the capabilities that Julia is designed with.
Designed with beginners in mind, this workshop helps you make the most of Python libraries and the Jupyter Notebook's functionality to understand how data science can be applied to solve real-world data problems. Key Features Gain useful insights into data science and machine learning Explore the different functionalities and features of a Jupyter Notebook Discover how Python libraries are used with Jupyter for data analysis Book DescriptionFrom banking and manufacturing through to education and entertainment, using data science for business has revolutionized almost every sector in the modern world. It has an important role to play in everything from app development to network security. Taking an interactive approach to learning the fundamentals, this book is ideal for beginners. You'll learn all the best practices and techniques for applying data science in the context of real-world scenarios and examples. Starting with an introduction to data science and machine learning, you'll start by getting to grips with Jupyter functionality and features. You'll use Python libraries like sci-kit learn, pandas, Matplotlib, and Seaborn to perform data analysis and data preprocessing on real-world datasets from within your own Jupyter environment. Progressing through the chapters, you'll train classification models using sci-kit learn, and assess model performance using advanced validation techniques. Towards the end, you'll use Jupyter Notebooks to document your research, build stakeholder reports, and even analyze web performance data. By the end of The Applied Data Science Workshop, you'll be prepared to progress from being a beginner to taking your skills to the next level by confidently applying data science techniques and tools to real-world projects. What you will learn Understand the key opportunities and challenges in data science Use Jupyter for data science tasks such as data analysis and modeling Run exploratory data analysis within a Jupyter Notebook Visualize data with pairwise scatter plots and segmented distribution Assess model performance with advanced validation techniques Parse HTML responses and analyze HTTP requests Who this book is forIf you are an aspiring data scientist who wants to build a career in data science or a developer who wants to explore the applications of data science from scratch and analyze data in Jupyter using Python libraries, then this book is for you. Although a brief understanding of Python programming and machine learning is recommended to help you grasp the topics covered in the book more quickly, it is not mandatory.
Create and share livecode, equations, visualizations, and explanatory text, in both a single document and a web browser with Jupyter Key Features Learn how to use Jupyter 5.x features such as cell tagging and attractive table styles Leverage big data tools and datasets with different Python packages Explore multiple-user Jupyter Notebook servers Book DescriptionThe Jupyter Notebook allows you to create and share documents that contain live code, equations, visualizations, and explanatory text. The Jupyter Notebook system is extensively used in domains such as data cleaning and transformation, numerical simulation, statistical modeling, and machine learning. Learning Jupyter 5 will help you get to grips with interactive computing using real-world examples. The book starts with a detailed overview of the Jupyter Notebook system and its installation in different environments. Next, you will learn to integrate the Jupyter system with different programming languages such as R, Python, Java, JavaScript, and Julia, and explore various versions and packages that are compatible with the Notebook system. Moving ahead, you will master interactive widgets and namespaces and work with Jupyter in a multi-user mode. By the end of this book, you will have used Jupyter with a big dataset and be able to apply all the functionalities you've explored throughout the book. You will also have learned all about the Jupyter Notebook and be able to start performing data transformation, numerical simulation, and data visualization. What you will learn Install and run the Jupyter Notebook system on your machine Implement programming languages such as R, Python, Julia, and JavaScript with the Jupyter Notebook Use interactive widgets to manipulate and visualize data in real time Start sharing your Notebook with colleagues Invite your colleagues to work with you on the same Notebook Organize your Notebook using Jupyter namespaces Access big data in Jupyter for dealing with large datasets using Spark Who this book is forLearning Jupyter 5 is for developers, data scientists, machine learning users, and anyone working on data analysis or data science projects across different teams. Data science professionals will also find this book useful for performing technical and scientific computing collaboratively.
A problem-solution guide to encounter various NLP tasks utilizing Java open source libraries and cloud-based solutions Key Features Perform simple-to-complex NLP text processing tasks using modern Java libraries Extract relationships between different text complexities using a problem-solution approach Utilize cloud-based APIs to perform machine translation operations Book DescriptionNatural Language Processing (NLP) has become one of the prime technologies for processing very large amounts of unstructured data from disparate information sources. This book includes a wide set of recipes and quick methods that solve challenges in text syntax, semantics, and speech tasks. At the beginning of the book, you'll learn important NLP techniques, such as identifying parts of speech, tagging words, and analyzing word semantics. You will learn how to perform lexical analysis and use machine learning techniques to speed up NLP operations. With independent recipes, you will explore techniques for customizing your existing NLP engines/models using Java libraries such as OpenNLP and the Stanford NLP library. You will also learn how to use NLP processing features from cloud-based sources, including Google and Amazon's AWS. You will master core tasks, such as stemming, lemmatization, part-of-speech tagging, and named entity recognition. You will also learn about sentiment analysis, semantic text similarity, language identification, machine translation, and text summarization. By the end of this book, you will be ready to become a professional NLP expert using a problem-solution approach to analyze any sort of text, sentences, or semantic words. What you will learn Explore how to use tokenizers in NLP processing Implement NLP techniques in machine learning and deep learning applications Identify sentences within the text and learn how to train specialized NER models Learn how to classify documents and perform sentiment analysis Find semantic similarities between text elements and extract text from a variety of sources Preprocess text from a variety of data sources Learn how to identify and translate languages Who this book is forThis book is for data scientists, NLP engineers, and machine learning developers who want to perform their work on linguistic applications faster with the use of popular libraries on JVM machines. This book will help you build real-world NLP applications using a recipe-based approach. Prior knowledge of Natural Language Processing basics and Java programming is expected. |
You may like...
Data Structures and Other Objects Using…
Michael Main, Walter J. Savitch
Paperback
R5,028
Discovery Miles 50 280
Data Abstraction and Problem Solving…
Janet Prichard, Frank Carrano
Paperback
R2,280
Discovery Miles 22 800
How To Pastel - Your Step By Step Guide…
Howexpert, Farcas Fabiana
Hardcover
R760
Discovery Miles 7 600
|