Welcome to Kubernetes. We are excited about the prospect of you joining our community! The Kubernetes community abides by the CNCF code of conduct. Here is an excerpt:
As contributors and maintainers of this project, and in the interest of fostering an open and welcoming community, we pledge to respect all people who contribute through reporting issues, posting feature requests, updating documentation, submitting pull requests or patches, and other activities.
HNC is a small project, and we have limited abilities to help onboard developers. If you'd like to contribute to the core of HNC, it would be helpful if you've created your own controllers before using controller-runtime and have a good understanding at least one non-functional task such as monitoring or lifecycle management. However, there are sometimes tasks to help improve the CLI or other aspects of usability that require less background knowledge.
With that said, if you want to use HNC yourself and are also a developer, we want to know what does and does not work for you, and we'd welcome any PRs that might solve your problems.
The main design doc is here; other design docs are listed here.
We have full documentation on how to get started contributing here:
- Sign the Contributor License Agreement Kubernetes projects require that you sign a Contributor License Agreement (CLA) before we can accept your pull requests. If you forget to do this, no worries, you'll be reminded when you make your first PR.
- Review the Kubernetes Contributor Guide - or you can just jump directly to the contributing section
- Glance over the Contributor Cheat Sheet
These last two docs are generall for all of Kubernetes and may not always be applicable to HNC. But you should at least glance over them.
Make sure you have installed the following libraries/packages and that they're
accessible from your PATH
:
- Docker
- kubectl
- kustomize
- kubebuilder
- This Github issue may help you resolve errors you get when running the tests or any other command.
If you're using gcloud
and the GCP Container Registry, make sure that gcloud
is configured to use the project containing the registry you want to use, and
that you've previously run gcloud auth configure-docker
so that Docker can use
your GCP credentials.
If you are developing on MacOS, you will need gnu-sed. Instructions are below:
brew install gnu-sed
ln -s $(which gsed) /usr/local/bin/sed # sym link sed to gsed
Once HNC is installed via make deploy
(see next sections), the development
cycle looks like the following:
- Start a new branch and make all your changes there (this is just the standard Github flow).
- Make changes locally and write new unit and e2e tests as necessary
- Unit tests are located in the same directory as the code, and either test
the code directly (e.g. in
internal/object
) or use the K8s test env, which is basically just an apiserver (e.g. ininternal/reconcilers
). - End-to-end (e2e) tests are located in
test/e2e/
, or other more special-purpose directories intest/
. These are basically glorifiedbash
scripts with better error checking, and directly invokekubectl
.
- Unit tests are located in the same directory as the code, and either test
the code directly (e.g. in
- Ensure
make test
passes - this runs unit tests only.- If you see all tests fail with
no matches for v1/CRD
error , remove your/usr/local/kubebuilder/
directory and reinstall kubebuilder.
- If you see all tests fail with
- Deploy to your cluster with
make deploy
- Test your changes by hand and verify that your changes are working
correctly. Some ways you can do that are:
- Look at logging with
make deploy-watch
- Look at the result of the structure of your namespaces with
kubectl-hns tree -A
orkubectl-hns tree NAMESPACE
- See the resultant conditions or labels on namespaces by using
kubectl describe namespace NAMESPACE
- Look at logging with
- Run the e2e tests via
make test-e2e
. This will take about 10-20m, so go get a coffee or walk around your house if you're quarantined.- Note that the deployment must be ready before you start the tests.
- You need to set the
HNC_REPAIR
env var to point to the absolute path of the manifests used to deploy HNC - either the released versions (e.g. stored on Github) or the full path to the local manifests. If these are not set, we'll skip any tests that include modifying the HNC deployement, e.g. to bypass webhooks.
- Once you're ready to make a pull request, please follow the following
instructions:
- Each PR should contain one commit. If you have multiple commits (either introduced during your development or as a result of comments during the PR review process), please squash them into a single commit. Force-push any changes to your fork of this repo.
- Ensure your commit message includes a "Tested:" section explaining in reasonable detail what you did to test your change. Here is a good example. A minimal message might be something like "Added new test; verified that the test failed before my change and passed after it; ran e2e tests."
- Create the PR. We'll try to reply fairly quickly!
- Make any requested changes (using
git commit --amend
andgit push -f
to ensure you only have one commit). - Profit! Or at least, enjoy your feeling of accomplishment.
If you need to make a larger change, please write your plan down somewhere - for example, either in a Github issue or in a short Google doc example.
To build from source and deploy to a cluster:
- Ensure your
kubeconfig
is configured to point at your cluster- On GKE, run
gcloud container clusters get-credentials <cluster-name> --zone <cluster-zone>
. Also ensure you rungcloud auth configure-docker
so thatdocker-push
works correctly. - To deploy to KIND, see below instead.
- On other platforms... do whatever it is you're supposed to do (and update this documentation with more useful guidance!).
- On GKE, run
- Use
make deploy
to deploy to your cluster.- This will also install the
kubectl-hns
plugin into$GOPATH/bin
. Ensure that this is in yourPATH
env var if you want to use it by sayingkubectl hns
, as described in the user guide. - The manifests that get deployed will be output to
/manifests/hnc-controller.yaml
if you want to see exactly what gets deployed. - Note that
make deploy
can respond to env vars in your environment; see the Makefile for more information.
- This will also install the
- To view logs, say
make deploy-watch
While developing the HNC, it's usually faster to deploy locally to KIND. But be aware of the following gotchas:
- When you install KIND, make sure you're not in the HNC directory, otherwise all kinds of Go module stuff will get messed up (this is a general KIND issue).
- Instead of
make deploy
, saymake kind-deploy
(or, equivalently,CONFIG=kind make deploy
ormake deploy CONFIG=kind
). This bypasses the steps where you try to push the image to a Docker registry like Docker Hub or GCP Container Registry (gcr.io
). - It's up to you to ensure that your kubectl context is pointed to your KIND
cluster (use
kubectl config get-contexts
to see if it is). - Run
make kind-reset
to stop any existing KIND cluster and setup a new one. You don't need to run this every time, only when you're first starting development or you think your KIND cluster is in a bad state.
In addition, KIND doesn't integrate with any identity providers - that is, you
can't add "[email protected]" as a "regular user." So you'll have to use service
accounts and impersonate them to test things like RBAC rules. Use kubectl --as system:serviceaccount:<namespace>:<sa-name>
to impersonate a service account
from the command line, as documented
here.
We think you can also use client certificate requests but we haven't tried that out. If you don't know what we're talking about, you probably don't want to try it out either. If you do, please update these docs when you get the chance!
Alternatively, you can also run the controller locally (ie, not on the cluster)
by saying make run
. You still need a cluster to exist (and your kubeconfig
must be pointing to it) but it's sometimes nice to have everything running on
your machine - e.g., if you want to run a debugger.
Webhooks don't work in this mode because I haven't bothered to find an easy way to make them work yet. We don't really use this method ourselves anymore so it may not always work.
The directory structure is fairly standard for a Kubernetes project. The most interesting directories are probably:
/api
: the API definition./cmd
: top-level executables. Currently the manager and the kubectl plugin./hack
: various release scripts, end-to-end bash script tests and other miscellaneous files./internal/reconcilers
: the reconcilers and their tests/internal/validators
: validating admission controllers/internal/forest
: the in-memory data structure, shared between the reconcilers and validators/internal/kubectl
: implementation of thekubectl-hns
plugin/test
: various end-to-end tests
Within the reconcilers
directory, there are four reconcilers:
- HNCConfiguration reconciler: manages the HNCConfiguration via the
cluster-wide
config
singleton. - Anchor reconciler: manages the subnamespace anchors via
the
subnamespaceanchor
resources. - HierarchyConfiguration reconciler: manages the hierarchy and the
namespaces via the
hierarchy
singleton per namespace. - Object reconciler: propagates (copies and deletes) the relevant objects
from parents to children. Instantiated once for every supported object GVK
(group/version/kind) - e.g.,
Role
,Secret
, etc.
HNC is overseen by the Working Group on Multi-Tenancy (wg-multitenancy). Please join us on Slack, mailing lists, and at our meeting at our community page.