spark-operator

Developer Guide

Build the Operator

In case you want to build the operator from the source code, e.g., to test a fix or a feature you write, you can do so following the instructions below.

The easiest way to build the operator without worrying about its dependencies is to just build an image using the Dockerfile.

$ docker build -t <image-tag> .

The operator image is built upon a base Spark image that defaults to gcr.io/spark-operator/spark:v3.0.0. If you want to use your own Spark image (e.g., an image with a different version of Spark or some custom dependencies), specify the argument SPARK_IMAGE as the following example shows:

$ docker build --build-arg SPARK_IMAGE=<your Spark image> -t <image-tag> .

If you want to use the operator on OpenShift clusters, first make sure you have Docker version 18.09.3 or above, then build your operator image using the OpenShift-specific Dockerfile.

$ export DOCKER_BUILDKIT=1
$ docker build -t <image-tag> -f Dockerfile.rh .

If you’d like to build/test the spark-operator locally, follow the instructions below:

$ mkdir -p $GOPATH/src/github.com/GoogleCloudPlatform
$ cd $GOPATH/src/github.com/GoogleCloudPlatform
$ git clone git@github.com:GoogleCloudPlatform/spark-on-k8s-operator.git
$ cd spark-on-k8s-operator

To update the auto-generated code, run the following command. (This step is only required if the CRD types have been changed):

$ hack/update-codegen.sh

To update the auto-generated CRD definitions, run the following command:

$ GO111MODULE=off go get -u sigs.k8s.io/controller-tools/cmd/controller-gen
$ controller-gen crd:trivialVersions=true,maxDescLen=0 paths="./pkg/apis/sparkoperator.k8s.io/v1beta2" output:crd:artifacts:config=./manifest/crds/

You can verify the current auto-generated code is up to date with:

$ hack/verify-codegen.sh

To build the operator, run the following command:

$ GOOS=linux go build -o spark-operator

To run unit tests, run the following command:

$ go test ./...

Build the API Specification Doc

When you update the API, or specifically the SparkApplication and ScheduledSparkApplication specifications, the API specification doc needs to be updated. To update the API specification doc, run the following command:

make build-api-docs

Running the above command will update the file docs/api-docs.md.