39,59 €
Java is the most popular programming language, according to the TIOBE index, and it is a typical choice for running production systems in many companies, both in the startup world and among large enterprises.
Not surprisingly, it is also a common choice for creating data science applications: it is fast and has a great set of data processing tools, both built-in and external. What is more, choosing Java for data science allows you to easily integrate solutions with existing software, and bring data science into production with less effort.
This book will teach you how to create data science applications with Java. First, we will revise the most important things when starting a data science application, and then brush up the basics of Java and machine learning before diving into more advanced topics. We start by going over the existing libraries for data processing and libraries with machine learning algorithms. After that, we cover topics such as classification and regression, dimensionality reduction and clustering, information retrieval and natural language processing, and deep learning and big data.
Finally, we finish the book by talking about the ways to deploy the model and evaluate it in production settings.
Das E-Book können Sie in Legimi-Apps oder einer beliebigen App lesen, die das folgende Format unterstützen:
Seitenzahl: 438
Veröffentlichungsjahr: 2017
BIRMINGHAM - MUMBAI
Copyright © 2017 Packt Publishing
All rights reserved. No part of this book may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, without the prior written permission of the publisher, except in the case of brief quotations embedded in critical articles or reviews.
Every effort has been made in the preparation of this book to ensure the accuracy of the information presented. However, the information contained in this book is sold without warranty, either express or implied. Neither the author, nor Packt Publishing, and its dealers and distributors will be held liable for any damages caused or alleged to be caused directly or indirectly by this book.
Packt Publishing has endeavored to provide trademark information about all of the companies and products mentioned in this book by the appropriate use of capitals. However, Packt Publishing cannot guarantee the accuracy of this information.
First published: April 2017
Production reference: 1250417
ISBN 978-1-78217-427-1
www.packtpub.com
Author
Alexey Grigorev
Copy Editor
Laxmi Subramanian
Reviewers
Stanislav Bashkyrtsev Luca Massaron Prashant Verma
Project Coordinator
Shweta H Birwatkar
Commissioning Editor
Veena Pagare
Proofreader
Safis Editing
Acquisition Editor
Manish Nainani
Indexer
Aishwarya Gangawane
ContentDevelopmentEditor
Amrita Noronha
Graphics
Tania Dutta
Technical Editor Akash Patel
Deepti Tuscano
Production Coordinator
Nilesh Mohite
Alexey Grigorev is a skilled data scientist, machine learning engineer, and software developer with more than 7 years of professional experience.
He started his career as a Java developer working at a number of large and small companies, but after a while he switched to data science. Right now, Alexey works as a data scientist at Searchmetrics, where, in his day-to-day job, he actively uses Java and Python for data cleaning, data analysis, and modeling.
His areas of expertise are machine learning and text mining, but he also enjoys working on a broad set of problems, which is why he often participates in data science competitions on platforms such as kaggle.com.
You can connect with Alexey on LinkedIn at https://de.linkedin.com/in/agrigorev.
I would like to thank my wife, Larisa, and my son, Arkadij, for their patience and support while I was working on the book.
Stanislav Bashkyrtsev has been working with Java for the last 9 years. Last years were focused on automation and optimization of development processes. Luca Massaron is a data scientist and a marketing research director specialized in multivariate statistical analysis, machine learning, and customer insight with over a decade of experience in solving real-world problems and in generating value for stakeholders by applying reasoning, statistics, data mining, and algorithms. From being a pioneer of Web audience analysis in Italy to achieving the rank of top ten Kaggler, he has always been passionate about everything regarding data and analysis and about demonstrating the potentiality of data-driven knowledge discovery to both experts and nonexperts. Favoring simplicity over unnecessary sophistication, he believes that a lot can be achieved in data science just by doing the essential. He is the coauthor of five recently published books and he is just working on the sixth. For Packt Publishing he contributed as an author to Python Data Science Essentials (both 1st and 2nd editions), Regression Analysis with Python, and Large Scale Machine Learning with Python.
You can find him on LinkedIn at https://it.linkedin.com/in/lmassaron.
Prashant Verma started his IT carrier in 2011 as a Java developer in Ericsson working in telecom domain. After a couple of years of JAVA EE experience, he moved into big data domain, and has worked on almost all the popular big data technologies such as Hadoop, Spark, Flume, Mongo, Cassandra, and so on. He has also played with Scala. Currently, he works with QA Infotech as lead data engineer, working on solving e-learning domain problems using analytics and machine learning.
Prashant has worked for many companies such as Ericsson and QA Infotech, with domain knowledge of telecom and e-learning. Prashant has also been working as a freelance consultant in his free time.
I want to thank Packt Publishing for giving me the chance to review the book as well as my employer and my family for their patience while I was busy working on this book.
For support files and downloads related to your book, please visit www.PacktPub.com.
Did you know that Packt offers eBook versions of every book published, with PDF and ePub files available? You can upgrade to the eBook version at www.PacktPub.com and as a print book customer, you are entitled to a discount on the eBook copy. Get in touch with us at [email protected] for more details.
At www.PacktPub.com, you can also read a collection of free technical articles, sign up for a range of free newsletters and receive exclusive discounts and offers on Packt books and eBooks.
https://www.packtpub.com/mapt
Get the most in-demand software skills with Mapt. Mapt gives you full access to all Packt books and video courses, as well as industry-leading tools to help you plan your personal development and advance your career.
Fully searchable across every book published by Packt
Copy and paste, print, and bookmark content
On demand and accessible via a web browser
Thanks for purchasing this Packt book. At Packt, quality is at the heart of our editorial process. To help us improve, please leave us an honest review on this book's Amazon page at https://www.amazon.com/dp/1782174273.
If you'd like to join our team of regular reviewers, you can e-mail us at [email protected]. We award our regular reviewers with free eBooks and videos in exchange for their valuable feedback. Help us be relentless in improving our products!
Preface
What this book covers
What you need for this book
Who this book is for
Conventions
Reader feedback
Customer support
Downloading the example code
Downloading the color images of this book
Errata
Piracy
Questions
Data Science Using Java
Data science
Machine learning
Supervised learning
Unsupervised learning
Clustering
Dimensionality reduction
Natural Language Processing
Data science process models
CRISP-DM
A running example
Data science in Java
Data science libraries
Data processing libraries
Math and stats libraries
Machine learning and data mining libraries
Text processing
Summary
Data Processing Toolbox
Standard Java library
Collections
Input/Output
Reading input data
Writing ouput data
Streaming API
Extensions to the standard library
Apache Commons
Commons Lang
Commons IO
Commons Collections
Other commons modules
Google Guava
AOL Cyclops React
Accessing data
Text data and CSV
Web and HTML
JSON
Databases
DataFrames
Search engine - preparing data
Summary
Exploratory Data Analysis
Exploratory data analysis in Java
Search engine datasets
Apache Commons Math
Joinery
Interactive Exploratory Data Analysis in Java
JVM languages
Interactive Java
Joinery shell
Summary
Supervised Learning - Classification and Regression
Classification
Binary classification models
Smile
JSAT
LIBSVM and LIBLINEAR
Encog
Evaluation
Accuracy
Precision, recall, and F1
ROC and AU ROC (AUC)
Result validation
K-fold cross-validation
Training, validation, and testing
Case study - page prediction
Regression
Machine learning libraries for regression
Smile
JSAT
Other libraries
Evaluation
MSE
MAE
Case study - hardware performance
Summary
Unsupervised Learning - Clustering and Dimensionality Reduction
Dimensionality reduction
Unsupervised dimensionality reduction
Principal Component Analysis
Truncated SVD
Truncated SVD for categorical and sparse data
Random projection
Cluster analysis
Hierarchical methods
K-means
Choosing K in K-Means
DBSCAN
Clustering for supervised learning
Clusters as features
Clustering as dimensionality reduction
Supervised learning via clustering
Evaluation
Manual evaluation
Supervised evaluation
Unsupervised Evaluation
Summary
Working with Text - Natural Language Processing and Information Retrieval
Natural Language Processing and information retrieval
Vector Space Model - Bag of Words and TF-IDF
Vector space model implementation
Indexing and Apache Lucene
Natural Language Processing tools
Stanford CoreNLP
Customizing Apache Lucene
Machine learning for texts
Unsupervised learning for texts
Latent Semantic Analysis
Text clustering
Word embeddings
Supervised learning for texts
Text classification
Learning to rank for information retrieval
Reranking with Lucene
Summary
Extreme Gradient Boosting
Gradient Boosting Machines and XGBoost
Installing XGBoost
XGBoost in practice
XGBoost for classification
Parameter tuning
Text features
Feature importance
XGBoost for regression
XGBoost for learning to rank
Summary
Deep Learning with DeepLearning4J
Neural Networks and DeepLearning4J
ND4J - N-dimensional arrays for Java
Neural networks in DeepLearning4J
Convolutional Neural Networks
Deep learning for cats versus dogs
Reading the data
Creating the model
Monitoring the performance
Data augmentation
Running DeepLearning4J on GPU
Summary
Scaling Data Science
Apache Hadoop
Hadoop MapReduce
Common Crawl
Apache Spark
Link prediction
Reading the DBLP graph
Extracting features from the graph
Node features
Negative sampling
Edge features
Link Prediction with MLlib and XGBoost
Link suggestion
Summary
Deploying Data Science Models
Microservices
Spring Boot
Search engine service
Online evaluation
A/B testing
Multi-armed bandits
Summary
Data science has become a quite important tool for organizations nowadays: they have collected large amounts of data, and to be able to put it into good use, they need data science--the discipline about methods for extracting knowledge from data. Every day more and more companies realize that they can benefit from data science and utilize the data that they produce more effectively and more profitably.
It is especially true for IT companies, they already have the systems and the infrastructure for generating and processing the data. These systems are often written in Java--the language of choice for many large and small companies across the world. It is not a surprise, Java offers a very solid and mature ecosystem of libraries that are time proven and reliable, so many people trust Java and use it for creating their applications.
Thus, it is also a natural choice for many data processing applications. Since the existing systems are already in Java, it makes sense to use the same technology stack for data science, and integrate the machine learning model directly in the application's production code base.
This book will cover exactly that. We will first see how we can utilize Java’s toolbox for processing small and large datasets, then look into doing initial exploration data analysis. Next, we will review the Java libraries that implement common Machine Learning models for classification, regression, clustering, and dimensionality reduction problems. Then we will get into more advanced techniques and discuss Information Retrieval and Natural Language Processing, XGBoost, deep learning, and large scale tools for processing big datasets such as Apache Hadoop and Apache Spark. Finally, we will also have a look at how to evaluate and deploy the produced models such that the other services can use them.
We hope you will enjoy the book. Happy reading!
Chapter 1, Data Science Using Java, provides the overview of the existing tools available in Java as well and introduces the methodology for approaching Data Science projects, CRISP-DM. In this chapter, we also introduce our running example, building a search engine.
Chapter 2, Data Processing Toolbox, reviews the standard Java library: the Collection API for storing the data in memory, the IO API for reading and writing the data, and the Streaming API for a convenient way of organizing data processing pipelines. We will look at the extensions to the standard libraries such as Apache Commons Lang, Apache Commons IO, Google Guava, and AOL Cyclops React. Then, we will cover most common ways of storing the data--text and CSV files, HTML, JSON, and SQL Databases, and discuss how we can get the data from these data sources. We will finish this chapter by talking about the ways we can collect the data for the running example--the search engine, and how we prepare the data for that.
Chapter 3, Exploratory Data Analysis, performs the initial analysis of data with Java: we look at how to calculate common statistics such as the minimal and maximal values, the average value, and the standard deviation. We also talk a bit about interactive analysis and see what are the tools that allow us to visually inspect the data before building models. For the illustration in this chapter, we use the data we collect for the search engine.
Chapter 4, Supervised Learning - Classification and Regression, starts with Machine Learning, and then looks at the models for performing supervised learning in Java. Among others, we look at how to use the following libraries--Smile, JSAT, LIBSVM, LIBLINEAR, and Encog, and we see how we can use these libraries to solve the classification and regression problems. We use two examples here, first, we use the search engine data for predicting whether a URL will appear on the first page of results or not, which we use for illustrating the classification problem. Second, we predict how much time it takes to multiply two matrices on certain hardware given its characteristics, and we illustrate the regression problem with this example.
Chapter 5, Unsupervised Learning – Clustering and Dimensionality Reduction, explores the methods for Dimensionality Reduction available in Java, and we will learn how to apply PCA and Random Projection to reduce the dimensionality of this data. This is illustrated with the hardware performance dataset from the previous chapter. We also look at different ways to cluster data, including Agglomerative Clustering, K-Means, and DBSCAN, and we use the dataset with customer complaints as an example.
Chapter 6, Working with Text – Natural Language Processing and Information Retrieval, looks at how to use text in Data Science applications, and we learn how to extract more useful features for our search engine. We also look at Apache Lucene, a library for full-text indexing and searching, and Stanford CoreNLP, a library for performing Natural Language Processing. Next, we look at how we can represent words as vectors, and we learn how to build such embeddings from co-occurrence matrices and how to use existing ones like GloVe. We also look at how we can use machine learning for texts, and we illustrate it with a sentiment analysis problem where we apply LIBLINEAR to classify if a review is positive or negative.
Chapter 7, Extreme Gradient Boosting, covers how to use XGBoost in Java and tries to apply it to two problems we had previously, classifying whether the URL appears on the first page and predicting the time to multiply two matrices. Additionally, we look at how to solve the learning-to-rank problem with XGBoost and again use our search engine example as illustration.
Chapter 8, Deep Learning with DeepLearning4j, covers Deep Neural Networks and DeepLearning4j, a library for building and training these networks in Java. In particular, we talk about Convolutional Neural Nets and see how we can use them for image recognition--predicting whether it is a picture of a dog or a cat. Additionally, we discuss data augmentation--the way to generate more data, and also mention how we can speed up the training using GPUs. We finish the chapter by describing how to rent a GPU server on Amazon AWS.
Chapter 9, Scaling Data Science, talks about big data tools available in Java, Apache Hadoop, and Apache Spark. We illustrate it by looking at how we can process Common Crawl--the copy of the Internet, and calculate TF-IDF of each document there. Additionally, we look at the graph processing tools available in Apache Spark and build a recommendation system for scientists, we recommend a coauthor for the next possible paper.
Chapter 10, Deploying Data Science Models, looks at how we can expose the models to the rest of the world in such a way they are usable. Here we cover Spring Boot and talk how we can use the search engine model we developed to rank the articles from Common Crawl. We finish by discussing the ways to evaluate the performance of the models in the online settings and talk about A/B tests and Multi-Armed Bandits.
You need to have any latest system with at least 2GB RAM and a Windows 7 /Ubuntu 14.04/Mac OS X operating system. Further, you will need to have Java 1.8.0 or above and Maven 3.0.0 or above installed.
This book is intended for software engineers who are comfortable with developing Java applications and are familiar with the basic concepts of data science. Additionally, it will also be useful for data scientists who do not yet know Java, but want or need to learn it.
Feedback from our readers is always welcome. Let us know what you think about this book-what you liked or disliked. Reader feedback is important for us as it helps us develop titles that you will really get the most out of.
To send us general feedback, simply e-mail [email protected], and mention the book's title in the subject of your message.
If there is a topic that you have expertise in and you are interested in either writing or contributing to a book, see our author guide at www.packtpub.com/authors.
Now that you are the proud owner of a Packt book, we have a number of things to help you to get the most from your purchase.
You can download the example code files for this book from your account at http://www.packtpub.com. If you purchased this book elsewhere, you can visit http://www.packtpub.com/support and register to have the files e-mailed directly to you.
You can download the code files by following these steps:
Log in or register to our website using your e-mail address and password.
Hover the mouse pointer on the
SUPPORT
tab at the top.
Click on
Code Downloads & Errata
.
Enter the name of the book in the
Search
box.
Select the book for which you're looking to download the code files.
Choose from the drop-down menu where you purchased this book from.
Click on
Code Download
.
Once the file is downloaded, please make sure that you unzip or extract the folder using the latest version of:
WinRAR / 7-Zip for Windows
Zipeg / iZip / UnRarX for Mac
7-Zip / PeaZip for Linux
The code bundle for the book is also hosted on GitHub at https://github.com/PacktPublishing/Mastering-Java-for-Data-Science. We also have other code bundles from our rich catalog of books and videos available at https://github.com/PacktPublishing/. Check them out!
We also provide you with a PDF file that has color images of the screenshots/diagrams used in this book. The color images will help you better understand the changes in the output. You can download this file from https://www.packtpub.com/sites/default/files/downloads/MasteringJavaforDataScience_ColorImages.pdf.
Although we have taken every care to ensure the accuracy of our content, mistakes do happen. If you find a mistake in one of our books-maybe a mistake in the text or the code-we would be grateful if you could report this to us. By doing so, you can save other readers from frustration and help us improve subsequent versions of this book. If you find any errata, please report them by visiting http://www.packtpub.com/submit-errata, selecting your book, clicking on the Errata Submission Form link, and entering the details of your errata. Once your errata are verified, your submission will be accepted and the errata will be uploaded to our website or added to any list of existing errata under the Errata section of that title.
To view the previously submitted errata, go to https://www.packtpub.com/books/content/support and enter the name of the book in the search field. The required information will appear under the Errata section.
Piracy of copyrighted material on the Internet is an ongoing problem across all media. At Packt, we take the protection of our copyright and licenses very seriously. If you come across any illegal copies of our works in any form on the Internet, please provide us with the location address or website name immediately so that we can pursue a remedy.
Please contact us at [email protected] with a link to the suspected pirated material.
We appreciate your help in protecting our authors and our ability to bring you valuable content.
If you have a problem with any aspect of this book, you can contact us at [email protected], and we will do our best to address the problem.
This book is about building data science applications using the Java language. In this book, we will cover all the aspects of implementing projects from data preparation to model deployment.
The readers of this book are assumed to have some previous exposure to Java and data science, and the book will help to take this knowledge to the next level. This means learning how to effectively tackle a specific data science problem and get the most out of the available data.
This is an introductory chapter where we will prepare the foundation for all the other chapters. Here we will cover the following topics:
What is machine learning and data science?
Cross Industry Standard Process for Data Mining
(
CRIPS-DM
), a methodology for doing data science projects
Machine learning libraries in Java for medium and large-scale data science applications
By the end of this chapter, you will know how to approach a data science project and what Java libraries to use to do that.
Data science is the discipline of extracting actionable knowledge from data of various forms. The name data science emerged quite recently--it was invented by DJ Patil and Jeff Hammerbacher and popularized in the article Data Scientist: The Sexiest Job of the 21st Century in 2012. But the discipline itself had existed before for quite a while and previously was known by other names such as data mining or predictive analytics. Data science, like its predecessors, is built on statistics and machine learning algorithms for knowledge extraction and model building.
The science part of the term data science is no coincidence--if we look up science, its definition can be summarized to systematic organization of knowledge in terms testable explanations and predictions. This is exactly what data scientists do, by extracting patterns from available data, they can make predictions about future unseen data, and they make sure the predictions are validated beforehand.
Nowadays, data science is used across many fields, including (but not limited to):
Banking
: Risk management (for example, credit scoring), fraud detection, trading
Insurance
: Claims management (for example, accelerating claim approval), risk and losses estimation, also fraud detection
Health care
: Predicting diseases (such as strokes, diabetes, cancer) and relapses
Retail
and
e-commerce
: Market basket analysis (identifying product that go well together), recommendation engines, product categorization, and personalized searches
This book covers the following practical use cases:
Predicting whether an URL is likely to appear on the first page of a search engine
Predicting how fast an operation will be completed given the hardware specifications
Ranking text documents for a search engine
Checking whether there is a cat or a dog on a picture
Recommending friends in a social network
Processing large-scale textual data on a cluster of computers
In all these cases, we will use data science to learn from data and use the learned knowledge to solve a particular business problem.
We will also use a running example throughout the book, building a search engine. We will use it to illustrate many data science concepts such as, supervised machine learning, dimensionality reduction, text mining, and learning to rank models.
Machine learning is a part of computer science, and it is at the core of data science. The data itself, especially in big volumes, is hardly useful, but inside it hides highly valuable patterns. With the help of machine learning, we can recognize these hidden patterns, extract them, and then apply the learned information to the new unseen items.
For example, given the image of an animal, a machine learning algorithm can say whether the picture is a dog or a cat; or, given the history of a bank client, it will say how likely the client is to default, that is, to fail to pay the debt.
Often, machine learning models are seen as black boxes that take in a data point and output a prediction for it. In this book, we will look at what is inside these black boxes and see how and when it is best to use them.
The typical problems that machine learning solves can be categorized in the following groups:
Supervised learning
: For each data point, we have a
label--
extra information that describes the outcome that we want to learn. In the cats versus dogs case, the data point is an image of the animal; the label describes whether it's a dog or a cat.
Unsupervised learning
: We only have raw data points and no label information is available. For example, we have a collection of e-mails and we would like to group them based on how similar they are. There is no explicit label associated with the e-mails, which makes this problem unsupervised.
Semi-supervised learning
: Labels are given only for a part of the data.
Reinforcement learning
: Instead of labels, we have a
reward
; something the model gets by interacting with the
environment
it runs in. Based on the reward, it can adapt and maximize it. For example, a model that learns how to play chess gets a positive reward each time it eats a figure of the opponent, and gets a negative reward each time it loses a figure; and the reward is proportional to the value of the figure.
As we discussed previously, for supervised learning we have some information attached to each data point, the label, and we can train a model to use it and to learn from it. For example, if we want to build a model that tells us whether there is a dog or a cat on a picture, then the picture is the data point and the information whether it is a dog or a cat is the label. Another example is predicting the price of a house--the description of a house is the data point, and the price is the label.
We can group the algorithms of supervised learning into classification and regression algorithms based on the nature of this information.
In classification problems, the labels come from some fixed finite set of classes, such as {cat, dog}, {default, not default}, or {office, food, entertainment, home}. Depending on the number of classes, the classification problem can be binary (only two possible classes) or multi-class (several classes).
Examples of classification algorithms are Naive Bayes, logistic regression, perceptron, Support Vector Machine (SVM), and many others. We will discuss classification algorithms in more detail in the first part of Chapter 4, Supervised Learning - Classification and Regression.
In regression problems, the labels are real numbers. For example, a person can have a salary in the range from $0 per year to several billions per year. Hence, predicting the salary is a regression problem.
Examples of regression algorithms are linear regression, LASSO, Support Vector Regression (SVR), and others. These algorithms will be described in more detail in the second part of Chapter 4, Supervised Learning - Classification and Regression.
Some of the supervised learning methods are universal and can be applied to both classification and regression problems. For example, decision trees, random forest, and other tree-based methods can tackle both types. We will discuss one such algorithm, gradient boosting machines in Chapter 7, Extreme Gradient Boosting.
Neural networks can also deal with both classification and regression problems, and we will talk about them in Chapter 8, Deep Learning with DeepLearning4J.
Unsupervised learning covers the cases where we have no labels available, but still want to find some patterns hidden in the data. There are several types of unsupervised learning, and we will look into cluster analysis, or clustering and unsupervised dimensionality reduction.
Typically, when people talk about unsupervised learning, they talk about cluster analysis or clustering. A cluster analysis algorithm takes a set of data points and tries to categorize them into groups such that similar items belong to the same group, and different items do not. There are many ways where it can be used, for example, in customer segmentation or text categorization.
Customer segmentation is an example of clustering. Given some description of customers, we try to put them into groups such that the customers in one group have similar profiles and behave in a similar way. This information can be used to understand what do the people in these groups want, and this can be used to target them with better advertisements and other promotional messages.
Another example is text categorization. Given a collection of texts, we would like to find common topics among these texts and arrange the texts according to these topics. For example, given a set of complaints in an e-commerce store, we may want to put ones that talk about similar things together, and this should help the users of the system navigate through the complaints easier.
Examples of cluster analysis algorithms are hierarchical clustering, k-means, density-based spatial clustering of applications with noise (DBSCAN), and many others. We will talk about clustering in detail in the first part of Chapter 5, Unsupervised Learning - Clustering and Dimensionality Reduction.
Another group of unsupervised learning algorithms is dimensionality reduction algorithms. This group of algorithms compresses the dataset, keeping only the most useful information. If our dataset has too much information, it can be hard for a machine learning algorithm to use all of it at the same time. It may just take too long for the algorithm to process all the data and we would like to compress the data, so processing it takes less time.
There are multiple algorithms that can reduce the dimensionality of the data, including Principal Component Analysis (PCA), Locally linear embedding, and t-SNE. All these algorithms are examples of unsupervised dimensionality reduction techniques.
Not all dimensionality reduction algorithms are unsupervised; some of them can use labels to reduce the dimensionality better. For example, many feature selection algorithms rely on labels to see what features are useful and what are not.
We will talk more about this in Chapter 5, Unsupervised Learning - Clustering and Dimensionality Reduction.
Processing natural language texts is very complex, they are not very well structured and require a lot of cleaning and normalizing. Yet the amount of textual information around us is tremendous: a lot of text data is generated every minute, and it is very hard to retrieve useful information from them. Using data science and machine learning is very helpful for text problems as well; they allow us to find the right text, process it, and extract the valuable bits of information.
There are multiple ways we can use the text information. One example is information retrieval, or, simply, text search--given a user query and a collection of documents, we want to find what are the most relevant documents in the corpus with respect to the query, and present them to the user. Other applications include sentiment analysis--predicting whether a product review is positive, neutral or negative, or grouping the reviews according to how they talk about the products.
We will talk more about information retrieval, Natural Language Processing (NLP) and working with texts in Chapter 6, Working with Text - Natural Language Processing and Information Retrieval. Additionally, we will see how to process large amounts of text data in Chapter 9, Scaling Data Science.
The methods we can use for machine learning and data science are very important. What is equally important is the the way we create them and then put them to use in production systems. Data science process models help us make it more organized and systematic, which is why we will talk about them next.
Applying data science is much more than just selecting a suitable machine learning algorithm and using it on the data. It is always good to keep in mind that machine learning is only a small part of the project; there are other parts such as understanding the problem, collecting the data, testing the solution and deploying to the production.
When working on any project, not just data science ones, it is beneficial to break it down into smaller manageable pieces and complete them one-by-one. For data science, there are best practices that describe how to do it the best way, and they are called process models. There are multiple models, including CRISP-DM and OSEMN.
In this chapter, CRISP-DM is explained as Obtain, Scrub, Explore, Model, and iNterpret (OSEMN), which is more suitable for data analysis tasks and addresses many important steps to a lesser extent.
Cross Industry Standard Process for Data Mining (CRISP-DM) is a process methodology for developing data mining applications. It was created before the term data science became popular, it's reliable and time-tested by several generations of analytics. These practices are still useful nowadays and describe the high-level steps of any analytical project quite well.
The CRISP-DM methodology breaks down a project into the following steps:
Business understanding
Data understanding
Data preparation
Modeling
Evaluation
Deployment
The methodology itself defines much more than just these steps, but typically knowing what the steps are and what happens at each step is enough for a successful data science project. Let's look at each of these steps separately.
The first step is Business Understanding. This step aims at learning what kinds of problems the business has and what they want to achieve by solving these problems. To be successful, a data science application must be useful for the business. The result of this step is the formulation of a problem which we want to solve and what is the desired outcome of the project.
The second step is Data Understanding. In this step, we try to find out what data can be used to solve the problem. We also need to find out if we already have the data; if not, we need to think how we can we get it. Depending on what data we find (or do not find), we may want to alter the original goal.
When the data is collected, we need to explore it. The process of reviewing the data is often called Exploratory Data Analysis and it is an integral part of any data science project. it helps to understand the processes that created the data, and can already suggest approaches for tackling the problem. The result of this step is the knowledge about which data sources are needed to solve the problem. We will talk more about this step in Chapter 3, Exploratory Data Analysis.
The third step of CRISP-DM is Data Preparation. For a dataset to be useful, it needs to be cleaned and transformed to a tabular form. The tabular form means that each row corresponds to exactly one observation. If our data is not in this shape, it cannot be used by most of the machine learning algorithms. Thus, we need to prepare the data such that it eventually can be converted to a matrix form and fed to a model.
Also, there could be different datasets that contain the needed information, and they may not be homogenous. What this means is that we need to convert these datasets to some common format, which can be read by the model.
This step also includes Feature Engineering--the process of creating features that are most informative for the problem and describe the data in the best way.
Many data scientists say that they spend most of their time on this step when building Data Science applications. We will talk about this step in Chapter 2,Data Processing Toolbox and throughout the book.
The fourth step is Modeling. In this step, the data is already in the right shape and we feed it to different Machine Learning algorithms. This step also includes parameter tuning, feature selection, and selecting the best model.
Evaluation of the quality of the models from the machine learning point of view happens during this step. The most important thing to check is the ability to generalize, and this is typically done via cross validation. In this step, we also may want to go back to the previous step and do extra cleaning and feature engineering. The outcome is a model that is potentially useful for solving the problem defined in Step 1.
The fifth step is Evaluation. It includes evaluating the model from the business perspective--not from the machine learning perspective. This means that we need to perform a critical review of the results so far and plan the next steps. Does the model achieve what we want? Additionally, some of the findings may lead to reconsidering the initial question. After this step, we can go to the deployment step or re-iterate the process.
The, final, sixth step is Model Deployment. During this step, the produced model is added to the production, so the result is the model integrated to the live system. We will cover this step in Chapter 10, Deploying Data Science Models.
Often, evaluation is hard because it is not always possible to say whether the model achieves the desired result or not. In these cases, the evaluation and deployment steps can be combined into one, the model is deployed and applied only to a part of users, and then the data for evaluating the model is collected. We will also briefly cover the ways of doing them, such as A/B testing and multi-armed bandits, in the last chapter of the book.
There will be many practical use cases throughout the book, sometimes a couple in each chapter. But we will also have a running example, building a search engine. This problem is interesting for a number of reasons:
It is fun
Business in almost any domain can benefit from a search engine
Many businesses already have text data; often it is not used effectively, and its use can be improved
Processing text requires a lot of effort, and it is useful to learn to do this effectively
We will try to keep it simple, yet, with this example, we will touch on all the technical parts of the data science process throughout the book:
Data Understanding
: Which data can be useful for the problem? How can we obtain this data?
Data Preparation
: Once the data is obtained, how can we process it? If it is HTML, how do we extract text from it? How do we extract individual sentences and words from the text?
Modeling
: Ranking documents by their relevance with respect to a query is a data science problem and we will discuss how it can be approached.
Evaluation
: The search engine can be tested to see if it is useful for solving the business problem or not.
Deployment
: Finally, the engine can be deployed as a REST service or integrated directly to the live system.
We will obtain and prepare the data in Chapter 2, Data Processing Toolbox, understand the data in Chapter 3, Exploratory Data Analysis, build simple models and evaluate them in Chapter 4, Supervised Machine Learning - Classification and Regression, look at how to process text in Chapter 6, Working with Text - Natural Language Processing and Information Retrieval, see how to apply it to millions of webpages in Chapter 9, Scaling Data Science, and, finally, learn how we can deploy it in Chapter 10, Deploying Data Science Models.
In this book, we will use Java for doing data science projects. Java might not seem a good choice for data science at first glance, unlike Python or R, it has fewer data science and machine learning libraries, it is more verbose and lacks interactivity. On the other hand, it has a lot of upsides as follows:
Java is a statically typed language, which makes it easier to maintain the code base and harder to make silly mistakes--the compiler can detect some of them.
The standard library for data processing is very rich, and there are even richer external libraries.
Java code is typically faster than the code in scripting languages that are usually used for data science (such as R or Python).
Maven, the de-facto standard for dependency management in the Java world, makes it very easy to add new libraries to the project and avoid version conflicts.
Most of big data frameworks for scalable data processing are written in either Java or JVM languages, such as Apache Hadoop, Apache Spark, or Apache Flink.
Very often production systems are written in Java and building models in other languages adds unnecessary levels of complexity. Creating the models in Java makes it easier to integrate them to the product.
Next, we will look at the data science libraries available in Java.
While there are not as many data science libraries in Java compared to R, there are quite a few. Additionally, it is often possible to use machine learning and data mining libraries written in other JVM languages, such as Scala, Groovy, or Clojure. Because these languages share the runtime environment, it makes it very easy to import libraries written in Scala and use them directly in Java code.
We can divide the libraries into the following categories:
Data processing libraries
Math and stats libraries
Machine learning and data mining libraries
Text processing libraries
Now we will see each of them in detail.
The standard Java library is very rich and offers a lot of tools for data processing, such as collections, I/O tools, data streams, and means of parallel task execution.
There are very powerful extensions to the standard library such as:
Google Guava (
https://github.com/google/guava
) and Apache Common Collections (
https://commons.apache.org/collections/
) for richer collections
Apache Commons IO (
https://commons.apache.org/io/
) for simplified I/O
AOL Cyclops-React (
https://github.com/aol/cyclops-react
) for richer functional-way parallel streaming
We will cover both the standard API for data processing and its extensions in Chapter 2, Data Processing Toolbox. In this book, we will use Maven for including external libraries such as Google Guava or Apache Commons IO. It is a dependency management tool and allows to specify the external dependencies with a few lines of XML code. For example, to add Google Guava, it is enough to declare the following dependency in pom.xml:
<dependency> <groupId>com.google.guava</groupId> <artifactId>guava</artifactId> <version>19.0</version> </dependency>
When we do it, Maven will go to the Maven Central repository and download the dependency of the specified version. The best way to find the dependency snippets for pom.xml (such as the previous one) is to use the search at https://mvnrepository.com or your favorite search engine.
Java gives an easy way to access databases through Java Database Connectivity (JDBC)--a unified database access protocol. JDBC makes it possible to connect virtually any relational database that supports SQL, such as MySQL, MS SQL, Oracle, PostgreSQL, and many others. This allows moving the data manipulation from Java to the database side.
When it is not possible to use a database for handling tabular data, then we can use DataFrame libraries for doing it directly in Java. The DataFrame is a data structure that originally comes from R and it allows to easily manipulate textual data in the program, without resorting to external database.
For example, with DataFrames it is possible to filter rows based on some condition, apply the same operation to each element of a column, group by some condition or join with another DataFrame. Additionally, some data frame libraries make it easy to convert tabular data to a matrix form so that the data can be used by machine learning algorithms.
There are a few data frame libraries available in Java. Some of them are as follows:
Joinery (
https://cardillo.github.io/joinery/
)
Tablesaw (
https://github.com/lwhite1/tablesaw
)
Saddle (
https://saddle.github.io/
) a data frame library for Scala
Apache Spark DataFrames (
http://spark.apache.org/
)
We will also cover databases and data frames in Chapter 2, Data Processing Toolbox and we will use DataFrames throughout the book.
There are more complex data processing libraries such as Spring Batch (http://projects.spring.io/spring-batch/). They allow creating complex data pipelines (called ETLs from Extract-Transform-Load) and manage their execution.
Additionally, there are libraries for distributed data processing such as:
Apache Hadoop (
http://hadoop.apache.org/
)
Apache Spark (
http://spark.apache.org/
)
Apache Flink (
https://flink.apache.org/
)
We will talk about distributed data processing in Chapter 9, Scaling Data Science.
The math support in the standard Java library is quite limited, and only includes methods such as log for computing the logarithm, exp for computing the exponent and other basic methods.
There are external libraries with richer support of mathematics. For example:
Apache Commons Math (
http://commons.apache.org/math/
) for statistics, optimization, and linear algebra
Apache Mahout (
http://mahout.apache.org/
) for linear algebra, also includes a module for distributed linear algebra and machine learning
JBlas
(
http://jblas.org/
)
optimized and very fast linear algebra package that uses the BLAS library
Also, many machine learning libraries come with some extra math functionality, often linear algebra, stats, and optimization.
There are quite a few machine learning and data mining libraries available for Java and other JVM languages. Some of them are as follows:
Weka (
http://www.cs.waikato.ac.nz/ml/weka/
) is probably the most famous data mining library in Java, contains a lot of algorithms and has many extensions.
JavaML (
http://java-ml.sourceforge.net/
) is quite an old and reliable ML library, but unfortunately not updated anymore
Smile (
http://haifengl.github.io/smile/
) is a promising ML library that is under active development at the moment and a lot of new methods are being added there.
JSAT (
https://github.com/EdwardRaff/JSAT
) contains quite an impressive list of machine learning algorithms.
H2O (
http://www.h2o.ai/
) is a framework for distributed ML written in Java, but is available for multiple languages, including Scala, R, and Python.
Apache Mahout (
http://mahout.apache.org/
) is used for in-core (one machine) and distributed machine learning. The Mahout Samsara framework allows writing the code in a framework-independent way and then executes it on Spark, Flink, or H2O.
There are several libraries that specialize solely on neural networks:
Encog (
http://www.heatonresearch.com/encog/
)
DeepLearning4j (
http://deeplearning4j.org/
)
We will cover some of these libraries throughout the book.
It is possible to do simple text processing using only the standard Java library with classes such as StringTokenizer, the java.text package, or the regular expressions.
In addition to that, there is a big variety of text processing frameworks available for Java as follows:
Apache Lucene (
https://lucene.apache.org/
) is a library that is used for information retrieval
Stanford CoreNLP (
http://stanfordnlp.github.io/CoreNLP/
)
Apache OpenNLP (
https://opennlp.apache.org/
)
LingPipe (
http://alias-i.com/lingpipe/
)
GATE (
https://gate.ac.uk/
)
MALLET (
http://mallet.cs.umass.edu/
)
Smile (
http://haifengl.github.io/smile/
) also has some algorithms for NLP
Most NLP libraries have very similar functionality and coverage of algorithms, which is why selecting which one to use is usually a matter of habit or taste. They all typically have tokenization, parsing, part-of-speech tagging, named entity recognition, and other algorithms for text processing. Some of them (such as StanfordNLP) support multiple languages, and some support only English.
We will cover some of these libraries in Chapter 6, Working with Text - Natural Language Processing and Information Retrival.
In this chapter, we briefly discussed data science and what role machine learning plays in it. Then we talked about doing a data science project, and what methodologies are useful for it. We discussed one of them, CRISP-DM, the steps it defines, how these steps are related and the outcome of each step.
Finally, we spoke about why doing a data science project in Java is a good idea, it is statically compiled, it's fast, and often the existing production systems already run in Java. We also mentioned libraries and frameworks one can use to successfully accomplish a data science project using the Java language.
With this foundation, we will now go to the most important (and most time-consuming) step in a data science project--Data Preparation.
In the previous chapter, we discussed the best practices for approaching data science problems. We looked at CRISP-DM, which is the methodology for dealing with data mining projects, and one of the first steps there is data preprocessing. In this chapter, we will take a closer look at how to do this in Java.
Specifically, we will cover the following topics:
Standard Java library
Extensions to the standard library
Reading data from different sources such as text, HTML, JSON, and databases
DataFrames for manipulating tabular data
In the end, we will put everything together to prepare the data for the search engine.
By the end of this chapter, you will be able to process data such that it can be used for machine learning and further analysis.
The standard Java library is very rich and offers a lot of tools for data manipulation, including:
Collections for organizing data in memory
I/O for reading and writing data
Streaming APIs for making data transformations easy
In this chapter, we will look at all these tools in detail.