How to use a Machine Learning Model to Make Predictions on Streaming Data using PySpark

Overview Streaming data is a thriving concept in the machine learning space Learn how to use a machine learning model (such as logistic regression) to make predictions on streaming data using PySpark We’ll cover the basics of Streaming Data and Spark Streaming, and then dive into the implementation part   Introduction Picture this – every second, more than 8,500 Tweets are sent, more than 900 photos are uploaded on Instagram, more than 4,200 Skype calls are made, more than 78,000 […]

Read more

Starting to develop in PySpark with Jupyter installed in a Big Data Cluster

Is not a secret that Data Science tools like Jupyter, Apache Zeppelin or the more recently launched Cloud Data Lab and Jupyter Lab are a must be known for the day by day work so How could be combined the power of easily developing models and the capacity of computation of a Big Data Cluster? Well in this article I will share very simple step to start using Jupyter notebooks for PySpark in a Data Proc Cluster in GCP. Final goal Prerequisites 1. Have a Google Cloud […]

Read more