Getting started with OpenFaaS on minikube
This is a guide on how to setup OpenFaaS — Serverless Functions Made Simple on Kubernetes 1.8 with minikube. minikube is a Kubernetes distribution which lets you run a Kubernetes cluster on your laptop — it’s available for Mac and Linux, but is most commonly used with MacOS.
This post is based upon our Official Kubernetes Deployment guide
Getting set up with Minikube
- Install the xhyve driver or VirtualBox which will be used to create a Linux VM for minikube. I find the VirtualBox option to be the most reliable.
- Setup minikube using the official docs
curl -sL cli.openfaas.com | sudo sh
- Install the
helmCLI via brew:
brew install kubernetes-helm
- Start minikube:
Docker Captain’s tip: look out for when Docker Inc starts to ship Kubernetes built-into Docker for Mac and Windows. This will let you use Kubernetes without having to install additional software.
Deploy OpenFaaS to minikube
- Create a service account for Helm’s server component (tiller):
kubectl -n kube-system create sa tiller \
&& kubectl create clusterrolebinding tiller \
--clusterrole cluster-admin \
tillerwhich is Helm’s server-side component:
helm init --skip-refresh --upgrade --service-account tiller
- Clone faas-netes (Kubernetes driver for OpenFaaS):
git clone https://github.com/openfaas/faas-netes && cd faas-netes
- Minikube is not configured for RBAC, so we’ll pass an additional flag to turn it off:
helm upgrade --install --debug --reset-values --set async=false --set rbac=false openfaas openfaas/
You’ll now see the OpenFaaS pods being installed on your minikube cluster. Type in
kubectl get pods to see them:
NAME READY STATUS RESTARTS AGE
alertmanager-6dbdcddfc4-fjmrf 1/1 Running 0 1m
faas-netesd-7b5b7d9d4-h9ftx 1/1 Running 0 1m
gateway-965d6676d-7xcv9 1/1 Running 0 1m
prometheus-64f9844488-t2mvn 1/1 Running 0 1m
The API Gateway contains a minimal UI for testing functions and exposes a RESTful API for function management. The faas-netesd daemon is a Kubernetes controller which connects to the Kubernetes API server to manage Services, Deployments and Secrets.
Prometheus and AlertManager work in tandem to enable auto-scaling of functions to match demand. Prometheus metrics give you operational oversight and allow you to build powerful dashboards.
Prometheus dashboard example:
OpenFaaS uses the Docker image format for the creation and versioning of functions which means that unlike many other FaaS projects you can use this in production to do:
- vulnerability scanning
- rolling upgrades
You can also deploy OpenFaaS to your existing production cluster and make use of spare capacity. The core services require around 10–30MB of RAM each.
A key advantage of OpenFaaS is that it works with the container orchestration platform’s API, which means we integrate natively with both Kubernetes and Docker Swarm. Also, since our functions are properly versioned within a Docker registry, we can scale our functions on demand without any additional latency associated with frameworks that build functions on demand.
Scaffold a new function
faas-cli new --lang python hello
This will create
hello.yml along with a
handler folder containing your handler.py file and requirements.txt for any pip modules you may need. You can edit these at any time without worrying about maintaining a Dockerfile — we do that and use best practices:
- multi-stage builds
- non-root users
- Official Docker Alpine Linux builds for the base (this is swappable)
Build your function
Your function will be built on your local machine and then pushed to a Docker registry. Let’s use the Docker Hub — just edit the
hello.yml file and enter your user account name:
Now invoke a build. You will need Docker on your local system.
faas-cli build -f hello.yml
Push the versioned Docker image which contains your function up to the Docker Hub. If you’ve not logged into the Docker hub then type in
docker login before carrying on.
faas-cli push -f hello.yml
Once you have multiple functions you can use the
--parallel=N flag to build or push with multiple cores at once. The CLI also supports options such as
Deploy and test your function
Now you can deploy your function, see it listed and invoke it. Each time you invoke the function we collect metrics which are made available through Prometheus.
$ export gw=http://$(minikube ip):31112
$ faas-cli deploy -f hello.yml --gateway $gw
No existing function to remove
You are given a standard route for invoking the function in the deployment message, but can also use the CLI to save on typing:
$ echo test | faas-cli invoke hello --gateway $gw
Now list the functions deployed and you will see the invocation count has gone up.
$ faas-cli list --gateway $gw
Function Invocations Replicas
hello 1 1
Note: this command also accepts a
--verbose flag for more information.
Since we are running OpenFaaS on a remote cluster (a Linux VM) we set up a
--gateway override environmental variable. This could also be a remote host on a cloud platform. The alternative is to update the gateway value in your .yml file.
Moving beyond minikube
Once you are familiar with OpenFaaS on minikube you can deploy to any Kubernetes cluster running on Linux hosts. Here’s an example of OpenFaaS running on Kubernetes on Google’s GKE platform by Stefan Prodan of WeaveWorks demonstrating the built-in auto-scaling capabilities:
We have dozens of guides and blog available to get you up and running on the “FaaS Train” — head over to GitHub and bookmark our guides:
I’ve included a short 15min video from Dockercon 2017 in Copenhagen where I gave an overview of Serverless and OpenFaaS at the Moby Summit.
Make sure to follow OpenFaaS on Twitter for all the latest Cool Hacks, demos and news.