Take a look at our Data Processing books. Shulph carries a great selection of Data Processing books, and we are always adding more.
Master the intricacies of Elasticsearch 7.0 and use it to create flexible and scalable search solutions Key Features Master the latest distributed search and analytics capabilities of Elasticsearch 7.0 Perform searching, indexing, and aggregation of your data at scale Discover tips and techniques for speeding up your search query performance Book Description Building enterprise-grade distributed applications and executing systematic search operations call for a strong understanding of Elasticsearch and expertise in using its core APIs and latest features. This book will help you master the advanced functionalities of Elasticsearch and understand how you can develop a sophisticated, real-time search engine confidently. In addition to this, you'll also learn to run machine learning jobs in Elasticsearch to speed up routine tasks. You'll get started by learning to use Elasticsearch features on Hadoop and Spark and make search results faster, thereby improving the speed of query results and enhancing the customer experience. You'll then get up to speed with performing analytics by building a metrics pipeline, defining queries, and using Kibana for intuitive visualizations that help provide decision-makers with better insights. The book will later guide you through using Logstash with examples to collect, parse, and enrich logs before indexing them in Elasticsearch. By the end of this book, you will have comprehensive knowledge of advanced topics such as Apache Spark support, machine learning using Elasticsearch and scikit-learn, and real-time analytics, along with the expertise you need to increase business productivity, perform analytics, and get the very best out of Elasticsearch. What you will learn Pre-process documents before indexing in ingest pipelines Learn how to model your data in the real world Get to grips with using Elasticsearch for exploratory data analysis Understand how to build analytics and RESTful services Use Kibana, Logstash, and Beats for dashboard applications Get up to speed with Spark and Elasticsearch for real-time analytics Explore the basics of Spring Data Elasticsearch, and understand how to index, search, and query in a Spring application Who this book is for This book is for Elasticsearch developers and data engineers who want to take their basic knowledge of Elasticsearch to the next level and use it to build enterprise-grade distributed search applications. Prior experience of working with Elasticsearch will be useful to get the most out of this book.
Build efficient data flow and machine learning programs with this flexible, multi-functional open-source cluster-computing framework Key Features Master the art of real-time big data processing and machine learning Explore a wide range of use-cases to analyze large data Discover ways to optimize your work by using many features of Spark 2.x and Scala Book Description Apache Spark is an in-memory, cluster-based data processing system that provides a wide range of functionalities such as big data processing, analytics, machine learning, and more. With this Learning Path, you can take your knowledge of Apache Spark to the next level by learning how to expand Spark's functionality and building your own data flow and machine learning programs on this platform. You will work with the different modules in Apache Spark, such as interactive querying with Spark SQL, using DataFrames and datasets, implementing streaming analytics with Spark Streaming, and applying machine learning and deep learning techniques on Spark using MLlib and various external tools. By the end of this elaborately designed Learning Path, you will have all the knowledge you need to master Apache Spark, and build your own big data processing and analytics pipeline quickly and without any hassle. This Learning Path includes content from the following Packt products: Mastering Apache Spark 2.x by Romeo Kienzler Scala and Spark for Big Data Analytics by Md. Rezaul Karim, Sridhar Alla Apache Spark 2.x Machine Learning Cookbook by Siamak Amirghodsi, Meenakshi Rajendran, Broderick Hall, Shuen MeiCookbook What you will learn Get to grips with all the features of Apache Spark 2.x Perform highly optimized real-time big data processing Use ML and DL techniques with Spark MLlib and third-party tools Analyze structured and unstructured data using SparkSQL and GraphX Understand tuning, debugging, and monitoring of big data applications Build scalable and fault-tolerant streaming applications Develop scalable recommendation engines Who this book is for If you are an intermediate-level Spark developer looking to master the advanced capabilities and use-cases of Apache Spark 2.x, this Learning Path is ideal for you. Big data professionals who want to learn how to integrate and use the features of Apache Spark and build a strong big data pipeline will also find this Learning Path useful. To grasp the concepts explained in this Learning Path, you must know the fundamentals of Apache Spark and Scala.
A solution-based guide to put your deep learning models into production with the power of Apache Spark Key Features Discover practical recipes for distributed deep learning with Apache Spark Learn to use libraries such as Keras and TensorFlow Solve problems in order to train your deep learning models on Apache Spark Book Description With deep learning gaining rapid mainstream adoption in modern-day industries, organizations are looking for ways to unite popular big data tools with highly efficient deep learning libraries. As a result, this will help deep learning models train with higher efficiency and speed. With the help of the Apache Spark Deep Learning Cookbook, you'll work through specific recipes to generate outcomes for deep learning algorithms, without getting bogged down in theory. From setting up Apache Spark for deep learning to implementing types of neural net, this book tackles both common and not so common problems to perform deep learning on a distributed environment. In addition to this, you'll get access to deep learning code within Spark that can be reused to answer similar problems or tweaked to answer slightly different problems. You will also learn how to stream and cluster your data with Spark. Once you have got to grips with the basics, you'll explore how to implement and deploy deep learning models, such as Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) in Spark, using popular libraries such as TensorFlow and Keras. By the end of the book, you'll have the expertise to train and deploy efficient deep learning models on Apache Spark. What you will learn Set up a fully functional Spark environment Understand practical machine learning and deep learning concepts Apply built-in machine learning libraries within Spark Explore libraries that are compatible with TensorFlow and Keras Explore NLP models such as Word2vec and TF-IDF on Spark Organize dataframes for deep learning evaluation Apply testing and training modeling to ensure accuracy Access readily available code that may be reusable Who this book is for If you're looking for a practical and highly useful resource for implementing efficiently distributed deep learning models with Apache Spark, then the Apache Spark Deep Learning Cookbook is for you. Knowledge of the core machine learning concepts and a basic understanding of the Apache Spark framework is required to get the best out of this book. Additionally, some programming knowledge in Python is a plus.
Understand the Blockchain revolution and get to grips with Ethereum, Hyperledger Fabric, and Corda.Key FeaturesResolve common challenges and problems faced in the Blockchain domainStudy architecture, concepts, terminologies, and DappsMake smart choices using Blockchain for personal and business investmentsBook DescriptionBlockchain Quick Reference takes you through the electrifying world of blockchain technology and is designed for those who want to polish their existing knowledge regarding the various pillars of the blockchain ecosystem.This book is your go-to guide, teaching you how to apply principles and ideas for making your life and business better. You will cover the architecture, Initial Coin Offerings (ICOs), tokens, smart contracts, and terminologies of the blockchain technology, before studying how they work. All you need is a curious mind to get started with blockchain technology. Once you have grasped the basics, you will explore components of Ethereum, such as ether tokens, transactions, and smart contracts, in order to build simple Dapps. You will then move on to learning why Solidity is used specifically for Ethereum-based projects, followed by exploring different types of blockchain with easy-to-follow examples. All this will help you tackle challenges and problems. By the end of this book, you will not only have solved current and future problems relating to blockchain technology but will also be able to build efficient decentralized applications.What you will learnUnderstand how blockchain architecture components workAcquaint yourself with cryptography and the mechanics behind blockchainApply consensus protocol to determine the business sustainabilityUnderstand what ICOs and crypto-mining are and how they workCreate cryptocurrency wallets and coins for transaction mechanismsUnderstand the use of Ethereum for smart contract and DApp developmentWho this book is forBlockchain Quick Reference is for you if you are a developer who wants to get well-versed with blockchain and its associated concepts and terminologies. You will explore the working mechanism of a decentralized application with the help of examples. Business leaders and blockchain enthusiasts will also find this book useful, as it will help you effectively address challenges and make better personal and business investments.
Gain hands-on experience with industry-standard data analysis and machine learning tools in Python Key Features Learn techniques to use data to identify the exact problem to be solved Visualize data using different graphs Identify how to select an appropriate algorithm for data extraction Book Description Data Science Projects with Python is designed to give you practical guidance on industry-standard data analysis and machine learning tools in Python, with the help of realistic data. The book will help you understand how you can use pandas and Matplotlib to critically examine a dataset with summary statistics and graphs, and extract the insights you seek to derive. You will continue to build on your knowledge as you learn how to prepare data and feed it to machine learning algorithms, such as regularized logistic regression and random forest, using the scikit-learn package. You'll discover how to tune the algorithms to provide the best predictions on new and, unseen data. As you delve into later chapters, you'll be able to understand the working and output of these algorithms and gain insight into not only the predictive capabilities of the models but also their reasons for making these predictions. By the end of this book, you will have the skills you need to confidently use various machine learning algorithms to perform detailed data analysis and extract meaningful insights from unstructured data. What you will learn Install the required packages to set up a data science coding environment Load data into a Jupyter Notebook running Python Use Matplotlib to create data visualizations Fit a model using scikit-learn Use lasso and ridge regression to reduce overfitting Fit and tune a random forest model and compare performance with logistic regression Create visuals using the output of the Jupyter Notebook Who this book is for If you are a data analyst, data scientist, or a business analyst who wants to get started with using Python and machine learning techniques to analyze data and predict outcomes, this book is for you. Basic knowledge of computer programming and data analytics is a must. Familiarity with mathematical concepts such as algebra and basic statistics will be useful.
Get the most out of Elasticsearch 7's new features to build, deploy, and manage efficient applications Key Features Discover the new features introduced in Elasticsearch 7 Explore techniques for distributed search, indexing, and clustering Gain hands-on knowledge of implementing Elasticsearch for your enterprise Book Description Elasticsearch is one of the most popular tools for distributed search and analytics. This Elasticsearch book highlights the latest features of Elasticsearch 7 and helps you understand how you can use them to build your own search applications with ease. Starting with an introduction to the Elastic Stack, this book will help you quickly get up to speed with using Elasticsearch. You'll learn how to install, configure, manage, secure, and deploy Elasticsearch clusters, as well as how to use your deployment to develop powerful search and analytics solutions. As you progress, you'll also understand how to troubleshoot any issues that you may encounter along the way. Finally, the book will help you explore the inner workings of Elasticsearch and gain insights into queries, analyzers, mappings, and aggregations as you learn to work with search results. By the end of this book, you'll have a basic understanding of how to build and deploy effective search and analytics solutions using Elasticsearch. What you will learn Install Elasticsearch and use it to safely store data and retrieve it when needed Work with a variety of analyzers and filters Discover techniques to improve search results in Elasticsearch Understand how to perform metric and bucket aggregations Implement best practices for moving clusters and applications to production Explore various techniques to secure your Elasticsearch clusters Who this book is for This book is for software developers, engineers, data architects, system administrators, and anyone who wants to get up and running with Elasticsearch 7. No prior experience with Elasticsearch is required.
Learn functional programming concepts and techniques to build Python applications. Key Features: - Study in detail all aspects of functional programming, including immutability, generators, and more. - Reinforce your learning through elaborate examples. - Learn how to implement advanced topics like closures, memoization, and monads in your applications. Book Description Python supports four programming paradigms – imperative, procedural, object-oriented, and functional. Of these, functional programming is probably the least understood and the least used. This book covers several topics that are directly and indirectly related to functional programming. After a quick overview of functional programming and its characteristics, Functional Programming in Python explains the various concepts of Python, starting with functions. You'll learn how to change the value of an object by using mutability. You'll then look at recursion as a more functional alternative to looping for certain algorithms, and learn how memoization alleviates the limitations of recursion in certain situations. The book further explains how to use closures as function factories and how to handle errors and exceptions with functors and monads. By the end of this book, you'll have all the knowledge you need for developing your applications with functional programming in Python. What you will learn: - Understand the advantages and disadvantages of functional programming - Use closures in your code to dynamically create functions - Create your own iterators - Use the general-purpose functools to create your own specialized reducing functions - Study and implement list and generator comprehensions to create lists - Create customized iterators with generators Who this book is for If you are a developer looking to create applications in Python using functional programming, this book is ideal for you. You will only need a basic knowledge of Python. Prior knowledge or experience of functional programming is not required.
Leverage the power of Tableau to get actionable business insights and make better business decisions Key Features Explore all the new features of Tableau 2018.x Begin your Tableau journey by understanding its core functionalities Apply advanced techniques such as data blending and parameters to make the best of Tableau Book Description Tableau is one of the leading business intelligence tools used worldwide, in organizations of every scale. In its latest release, Tableau 2018 promises richer and more useful features related to visual analytics, reporting, dashboarding, and a host of other data visualization aspects. Getting Started with Tableau 2018.x will get you up and running with these features. The book starts with all the new functionalities of the different Tableau 2018 versions, along with concrete examples of how to use them. However, if you're new to Tableau, don't worry! The rest of the book will guide you through each major aspect of Tableau with examples. You'll learn how to connect to data, build a data source, visualize your data, build a dashboard, and share it online. In the final chapters, you'll also learn advanced techniques such as creating a cross-database join, data blending, and more. By the end of the book, you will have a firm understanding of how to effectively use Tableau to create quick, cost-effective, and business-efficient business intelligence solutions. What you will learn Discover new functionalities such as density, extensions, and transparency introduced in Tableau 2018.x Connect tables and make transformations such as pivoting the field and splitting columns Build an efficient data source for your analysis Design insightful data visualization using different mark types and properties Build powerful dashboards and stories Share your work and interact with Tableau Server Use Tableau to explore your data and find new insights Explore Tableau's advanced features and gear up for upcoming challenges Who this book is for If you're a beginner or an existing BI professional looking to get the most out of Tableau 2018's latest features, this book is for you. You'll also find this book useful if you're an aspiring analyst who wants to learn the capabilities of Tableau to answer business-specific questions. No prior knowledge or experience with Tableau or any other Business Intelligence tool is necessary.