Spark Introduction
- Distributed System
- Hadoop, MapReduce and Spark
- Spark - RDD, DataFrame
- Pyspark
Steps to run PySpark and Jupyter Notebook using Docker
- Install Docker desktop app
- Create a new folder on your system
- Place the file docker-compose.yaml
- Run this file using command docker-compose up
- Open the URL http://127.0.0.1:8888/?token=YOUR_TOKEN