TensorFlow is now available on Apache Spark framework, but how do you get started? Read this !
Have you ever wanted to write an app that uses AI ? Follow along with me and you'll be there by the end of this short article.
Big Data is a huge topic that consists of many domains and expertise. All the way from DevOps, Data Engineers to Data Scientist, AI, Machine Learning, algorithm developers and many more. We all struggle with massive amounts of data. When we deal with a massive amount of data, we need the best minds and tools. This is where the magical combination of Apache Spark and Tensor Flow takes place and we call it TensorFrame.
Apache Spark took over the Big Data world, giving answers and supporting Data Engineers to be a more successful while, Data Scientist had to figure their way around the limitation of the machine learning library that Spark provides, the Spark MLlib.
But no more, now, there is TensorFlow support for Apache Spark users. These tools combined makes the work of Data Scientists more productive, more accurate and faster. And taking outcome from Research to Develop to Production, faster than ever.
Before we start, let's align the terms:
- Tensor Flow is an open source machine learning framework for high-performance numerical computations created by Google. It comes with strong support for AI: machine learning and deep learning.
- Azure Databricks is an Apache Spark-based analytics platform optimized for the Microsoft Azure cloud services platform. Azure Databricks also acts as Software as a Service( SaaS) / Big Data as a Service (BDaaS).
- TensorFrames is an Apache Spark component that enables us to create our own scalable TensorFlow learning algorithms on Spark Clusters.
-1- the workspace:
First, we need to create the workspace, we are using Databricks workspace and here is a tutorial for creating it.
-2- the cluster:
After we have the workspace, we need to create the cluster itself. Let’s create our spark cluster using this tutorial, make sure you have the next configurations in your cluster:
- A working version of Apache Spark (2.4 or greater)
- Java 8+
- (Optional) python 2.7+/3.6+ if you want to use the python interface.
- (Optional) the python TensorFlow package if you want to use the python interface. See the official instructions on how to get the latest release of TensorFlow.
- (Optional) pandas >= 0.19.1 if you want to use the python interface
with Databricks runtime versions or above :
Press start to launch the cluster
-3- import the library:
Under Azure Databricks, go to Common Tasks and click Import Library:
TensorFrame can be found on maven repository, so choose the Maven tag. Under Coordinates, insert the library of your choice, for now, it will be:
Click the Create button.
You will see this:
BOOM. you have TensorFrame on your Databricks Cluster.
-4- the notebook:
We use the notebook as our code notebook where we can write the code and run it directly on our Spark Cluster.
Now that we have a running cluster, let’s run a notebook:
Click the New Notebook and choose the programming language of your choice ( here we chose Scala)
This is how it looks like with scala code on the notebook portal with TensorFrames:
The code example is from Databricks git repository.
Now that you have everything up and running, what they will do with tensorflow?
Comment and let me know your thoughts. Your comments enable me to improve and get better every day. Thank you ❤️