All prow job configs for prow.k8s.io live here.
They are tested and validated by tests in config/tests
This document attempts to be a step-by-step, copy-pastable guide to the use of prow jobs for the Kubernetes project. It may fall out of date. For more info, it was sourced from the following:
There are three types of prow jobs:
- Presubmits run against code in PRs
- Postsubmits run after merging code
- Periodics run on a periodic basis
Please see ProwJob docs for more info
Where possible, we prefer that jobs use images that are pinned to a specific version, containing only what is needed.
Some good examples include:
- pull-release-unit uses
golang:1.12
to rungo test ./...
- pull-release-notes-lint uses
node:11
to runnpm ci && npm lint
- pull-org-test-all uses
launcher.gcr.io/google/bazel:0.26.0
to runbazel test //...
Many jobs use gcr.io/k8s-testimages/foo
images that are built from source in
images/
. Some of these have evolved organically, with way more dependencies
than needed, and will be periodically bumped by PRs. These are sources of
technical debt that are often not very well maintained. Use at your own risk,
eg:
- periodic-kubernetes-e2e-packages-pushed uses
gcr.io/k8s-staging-test-infra/kubekins:latest-master
to run./tests/e2e/packages/verify_packages_published.sh
which ends up runningapt-get
andyum
commands. Perhaps adebian
image would be better.
Prow supports Presets to define and patch in common
env vars and volumes used for credentials or common job config. Some are
defined centrally in [config/prow/config.yaml
], while others can be defined in
files here. eg:
preset-service-account: "true"
ensures the prowjob has a GCP service account in a well known location, with well known env vars pointing to it.preset-pull-kubernetes-e2e: "true"
sets environment variables to make kubernetes e2e tests less susceptible to flakespreset-aws-credentials: "true"
ensures the prowjob has AWS credentials for kops tests in a well known location, with an env var pointint to it- the default preset with no labels is used to set the
GOPROXY
env var for all jobs by default
Prow jobs can use secrets located in the same namespace within the cluster where the jobs are executed, by using the same mechanism of podspec. The secrets used in prow jobs can be source controlled and synced from any major secret manager provider, such as google secret manager, see prow_secret for instructions.
A presubmit job named "pull-community-verify" that will run against all PRs to
kubernetes/community's master branch. It will run make verify
in a checkout
of kubernetes/community at the PR's HEAD. It will report back to the PR via a
status context named pull-kubernetes-community
. Its logs and results are going
to end up in GCS under kubernetes-jenkins/pr-logs/pull/community
. Historical
results will display in testgrid on the sig-contribex-community
dashboard
under the pull-verify
tab
presubmits:
kubernetes/community:
- name: pull-community-verify # convention: (job type)-(repo name)-(suite name)
annotations:
testgrid-dashboards: sig-contribex-community
testgrid-tab-name: pull-verify
branches:
- master
decorate: true
always_run: true
spec:
containers:
- image: golang:1.12.5
command:
- /bin/bash
args:
- -c
# Add GOPATH/bin back to PATH to workaround #9469
- "export PATH=$GOPATH/bin:$PATH && make verify"
A periodic job named "periodic-cluster-api-provider-aws-test-creds" that will
run every 4 hours against kubernetes-sigs/cluster-api-provider-aws's master
branch. It will run ./scripts/ci-aws-cred-test.sh
in a checkout of the repo
located at sigs.k8s.io/cluster-api-provider-aws
. The presets it's using will
ensure it has aws credentials and aws ssh keys in well known locations. Its
logs and results are going to end up in GCS under
kubernetes-jenkins/logs/periodic-cluster-api-provider-aws-test-creds
.
Historical results will display in testgrid on the sig-cluster-lifecycle-cluster-api-provider-aws
dashboard under the test-creds
tab
It's using the kubekins-e2e
image which isn't recommended,
but works for now.
periodics:
- name: periodic-cluster-api-provider-aws-test-creds
annotations:
testgrid-dashboards: sig-cluster-lifecycle-cluster-api-provider-aws
testgrid-tab-name: test-creds
decorate: true
interval: 4h
labels:
preset-service-account: "true"
preset-aws-ssh: "true"
preset-aws-credential: "true"
extra_refs:
- org: kubernetes-sigs
repo: cluster-api-provider-aws
base_ref: master
path_alias: "sigs.k8s.io/cluster-api-provider-aws"
spec:
containers:
- image: gcr.io/k8s-staging-test-infra/kubekins-e2e:v20210917-ee1e7c845b-master
command:
- "./scripts/ci-aws-cred-test.sh"
- Find or create the prowjob config file in this directory
- In general jobs for
github.com/org/repo
useorg/repo/filename.yaml
- For kubernetes/kubernetes we prefer
kubernetes/sig-foo/filename.yaml
- Ensure
filename.yaml
is unique across the config subdir; prow uses this as a key in its configmap
- In general jobs for
- Ensure an
OWNERS
file exists in the directory for job, and has appropriate approvers/reviewers - Write or edit the job config (please see how-to-add-new-jobs)
- Ensure the job is configured to to display its results in testgrid.k8s.io
- The simple way: add testgrid annotations
- Please see the testgrid documentation for more details on configuation options
- Open a PR with the changes; when it merges @k8s-ci-robot will deploy the changes automatically
- Find the prowjob config file in this directory
- Remove the entry for your job; if that was the last job in the file, remove the file
- If the job had no testgrid annotations, ensure its [
testgrid/config.yaml
] entries are gone - Open a PR with the changes; when it merges @k8s-ci-robot will deploy the changes automatically
Please try using phaino
, it will interactively
help you run a docker command that approximates the pod that would be scheduled
on behalf of an actual prow job.
This requires a running instance of prow. In general we discourage the use of prow.k8s.io as a testbed for job development, and recommend the use of your own instance of prow for faster iteration. That said, an approach that people have used in the past with mostly-there jobs is to iterate via PRs; just recognize this is going to depend on review latency.
For more details, please refer to How to Test a ProwJob
Normally prow will automatically schedule your job, however if for some reason you need to trigger it again and are a Prow administrator you have a few options:
- you can use the rerun feature in prow.k8s.io to run the job again with the same config
- you can use
config/mkpj.sh
to create a prowjob CR from your local config - you can use
bazel run //prow/cmd/mkpj -- --job=foo ...
to create a prowjob CR from your local config
For the latter two options you'll need to submit the resulting CR via kubectl
configured against
the prow services cluster.
There are some sets of jobs that are generated and should not be edited by hand. These specific instructions should probably just live adjacent to the jobs rather than in this central README, but here we are for now.
These test different master/node image versions against multiple k8s branches. If you
want to change these, update releng/test_config.yaml
and then run
# from test-infra root
$ ./hack/update-generated-tests.sh
When a release branch of kubernetes is first cut, the current set of master jobs
must be forked to use the new release branch. Use releng/config-forker
to
accomplish this, eg:
# from test-infra root
$ bazel run //releng/config-forker -- \
--job-config $(pwd)/config/jobs \
--version 1.15 \
--output $(pwd)/config/jobs/kubernetes/sig-release/release-branch-jobs/1.15.yaml