A collection of useful Docker Compose configuration files, for speeding up tool deployment
For experimenting with PySpark on a Spark Cluster.
cd spark_cluster-pyspark_notebook
sudo docker-compose up --scale spark-worker=[NUMBER_OF_WORKERS]
(Fill in NUMBER_OF_WORKERS, int > 0)
- Access Jupyter Notebook via the link (with access token) found in terminal logs
- Access Spark GUI via http://localhost:8080/
- Volume is mounted on ./notebooks and ./data
Jupyter Notebook does not have write access on local volume (e.g. ./notebooks): sudo chmod 777 ./notebooks
Postgres database with PostGIS extension for geospatial analysis
Fill in database credentials in file database.env
cd postgis
sudo docker-compose up
- Access database via local port 8082 using your credentials