Overview
Mungegithub provides a number of tools intended to automate github processes. While mainly for the kubernetes community, some thought is put into making it generic. Mungegithub is built as a single binary, but is run in 3 different ways for 3 different purposes.
- submit-queue: This looks at open PRs and attempts to help automate the process getting PRs from open to merged.
- cherrypick: This looks at open and closed PRs with the
cherry-pick-candidate
label and attempts to help the branch managers deal with the cherry-pick process.
- shame-mailer: This looks at open issues and e-mails assignees who have not closed their issues rapidly.
One can see the specifics of how the submit-queue
and cherrypick
options are executed by looking at the deployment definition in their respective subdirectories.
One may also look in the example-one-off
directory for a small skeleton program which prints the number of all open PRs. It is an excellent place to start if you need to write a 'one-off' automation across a large number of PRs.
Building and running
Executing make help
inside the mungegithub
directory should inform you about the functions provided by the Makefile. A common pattern when developing is to run something like:
make mungegithub && ./mungegithub --dry-run --token-file=/path/to/token --once --www=submit-queue/www --kube-dir=$GOPATH/src/k8s.io/kubernetes --pr-mungers=submit-queue --min-pr-number=25000 --max-pr-number=25500 --organization=kubernetes --project=kubernetes --repo-dir=/tmp
A Github oauth token is required, even in readonly/test mode. For production, we use a token with write access to the repo in question. https://help.github.com/articles/creating-an-access-token-for-command-line-use/ discusses the procedure to get a personal oauth token. These tokens will need to be loaded into kubernetes secret
s for use by the submit and/or cherry-pick queue. It is extremely easy to use up the 5,000 API calls per hour so the production token should not be re-used for tests.
After successfully running the local binary one may build, test, and deploy in readonly mode to a real kube cluster. To do so, one must make sure that one's kubeconfig file is set up for the test/read-only cluster by running any necessary kubectl config
commands by hand. One may also need a container repository with read & write access. It is just as easy to create a new dockerhub account, and a public repository named submit-queue
within that. We will refer to the repository as docker.io/$USERNAME
where $USERNAME
is a placeholder. The steps required to deploy on a real cluster for the submit-queue application are as follows. The instructions to run the cherrypick application are along the same lines. Below, we explain the steps to run on the kubernetes main repository. Running on other repositories is similar, except that the corresponding YAML files are in a directory for that repository.
- Store your personal access token in a plain text file named (token) in the mungegithub directory.
- Run
APP=submit-queue; TARGET=<reponame>; make secret
to generate a local.secret.yaml.
- Run
kubectl --kubeconfig=... create -f mungegithub/submit-queue/deployment/<reponame>/local.secret.yaml
to load the secret.
- Run
kubectl --kubeconfig=... create -f mungegithub/submit-queue/deployment/<reponame>/pv.yaml
to create a persistent volume. (If you are running a local cluster, and not on GCP, use mungegithub/submit-queue/pv-local.yaml
to create a persistent volume on your host. The file may need to be modified to match the expected name of the persistent volume by the deployment).
- Run
kubectl --kubeconfig=... create -f mungegithub/submit-queue/deployment/<reponame>/pvc.yaml
to create a persistent volume claim.
- Check that the persistent volume claim is bound by checking
kubectl --kubeconfig=... get pvc
.
After these steps, we may need to push a configmap, in case any of the commandline arguments were changed. Pushing a new configmap for the kubernetes repository looks like the following:
TARGET=kubernetes APP=submit-queue KUBECONFIG=/path/to/kubeconfig make push_config
It can finally deployed as:
TARGET=kubernetes REPO=docker.io/$USERNAME APP=submit-queue KUBECONFIG=/path/to/kubeconfig make deploy
After this has successfully deployed to the test cluster in read-only mode, running in production involves running any required kubectl config
commands to point to the production cluster, pushing a configmap if necessary, and then running:
TARGET=kubernetes REPO=docker.io/$USERNAME APP=submit-queue KUBECONFIG=/path/to/kubeconfig READONLY=false make deploy
About the mungers
A small amount of information about some of the individual mungers inside each of the 3 varieties are listed below:
submit-queue
- block-paths - add
do-not-merge
label to PRs which change files which should not be changed (mainly old docs moved to kubernetes.github.io)
- blunderbuss - assigned PRs to individuals based on the contents of OWNERS files in the main repo
- cherrypick-auto-approve - adds
cherrypick-approved
to PRs in a release branch if the 'parent' pr in master was approved
- cherrypick-label-unapproved - adds
do-not-merge
label to PRs against a release-* branch which do not have cherrypick-approved
- comment-deleter - deletes comments created by the k8s-merge-robot which are no longer relevant. Such as comments about a rebase being required if it has been rebased.
- comment-deleter-jenkins - deleted comments create by the k8s-bot jenkins bot which are no longer relevant. Such as old test results.
- lgtm-after-commit - removes
lgtm
label if a PR is changed after the label was added
- needs-rebase - adds and removes a
needs-rebase
label if a PR needs to be rebased before it can be applied.
- path-label - adds labels, such as
kind/new-api
based on if ANY file which matches changed
- release-note-label - Manages the addition/removal of
release-note-label-required
and all of the rest of the release-note-*
labels.
- size - Adds the xs/s/m/l/xl labels and comments to PRs
- stale-green-ci - Reruns the CI tests every X hours (96?) for PRs which passed. So PRs which sit around for a long time will notice failures sooner.
- stale-pending-ci - Reruns the CI tests if they have been 'in progress'/'pending' for 24 hours.
- submit-queue - This is the brains that actually tracks and merges PRs. It also provides the web site interface.
cherrypick
- cherrypick-clear-after-merge - This watches for PRs against release branches which merged and removes the
cherrypick-candidate
label from the PR on master.
- cherrypick-must-have-milestone - This complains on any PR against a release branch which does not have a vX.Y milestone.
- cherrypick-queue - This is the web display of all PRs with the
cherrypick-candidate
label which a branch owner is likely to want to pay attention to.
Instructions on running mungegithub locally with your own repository
Sometimes we may want to run QA tests locally using the mungegithub binary. The steps to do this are as follows.
cd
to the contrib/mungegithub directory.
- Run
go build
to compile the mungegithub binary.
- Running the binary is as simple as running
./mungegithub
and supplying the appropriate flags.
- The flags that are essential are as follows:
--pr-mungers
, --organization
, --project
are required flags. Based on the mungers specified in pr-mungers, other flags may be required.
--token
or --token-file
are needed. It is highly recommended that you provide a GitHub access token without write access to the repositories you are running on, as an extra measure of safety.
- The
--dry-run=true
flag must be specified to ensure you're not posting comments accidentally.
- The
--repo-dir
should be pointed to /tmp if required.
- The
--www=submit-queue/www/
will start up the http server if specified with the submit-queue munger, and serve on localhost:8080.
Instructions on turning up a new submit-queue instance.
The steps below make use of the utility cluster which runs the existing submit-queues.
- Create a new directory for the repo on which you want to run the submit-queue instance. For example, if we want to call it
<TARGET>
, we create contrib/submit-queue/deployments/<TARGET>
.
- Add a service.yaml, pv.yaml, pvc.yaml, secret.yaml, configmap.yaml to the directory and configure them appropriately.
- The configmap’s name must be
<TARGET>-sq-flags
.
- The target-repo must be changed to
<TARGET>
.
- The PV and PVC must be named
<TARGET>-cache
.
- The secret must be named
<TARGET>-github-token
.
- The service must be named
<TARGET>-sq-status
.
- Create a persistent disk named
<TARGET>-cache
on the utility cluster. It is typically 10G in size.
- Switch context with kubectl to point to the utility cluster.
- Create the PV and PVC resources. After creation, the PV and PVC should be bound.
- Create a new secret using the below command, which uses an API token stored in
./token
, and generates a local.secret.yaml
file.
make secret APP=submit-queue TARGET=<TARGET>
- Load the secret created yaml using
kubectl create -f submit-queue/local.secret.yaml
.
- Create the service which is of type NodePort.
- Finally, update the ingress.yaml with the new URL and the new service to point to.
- Apply changes to the running ingress instance.
Communicating with the Bot
Github contributors and reviewers can communicate with the mungebot by commenting on a PR with the following commands entered alone in the text field. All commands require the following syntax: /<COMMAND> [OPTIONAL ARGS]
. Note the forward slashing preceding the command name.
List of Commands
- /lgtm : applies the lgtm label
- /lgtm cancel : removes a previously applied lgtm label