
# Specify the User that the actual main process will run asĮOF Containers: the hottest thing to hit 2009 Call to hacktion: adapting the Spark Dockerfileįirst, let’s download some Apache Spark release binaries and adapt the dockerfile so that it plays nicely with pyspark: wget Ĭhgrp root /etc/passwd & chmod ug+rw /etc/passwd & \ĬOPY kubernetes/dockerfiles/spark/entrypoint.sh /opt/ĬOPY kubernetes/dockerfiles/spark/decom.sh /opt/ Let’s take a look at getting Apache Spark on this thing so we can do all the data scientist stuff.


If so, you can exit any SSH session to your Ubuntu Core in the sky and return to your local system. If you’ve followed the steps in Part 1 and Part 2 of this series, you’ll have a working MicroK8s on the next-gen Ubuntu Core OS deployed, up, and running on the cloud with nested virtualisation using LXD.
