CAPM3 requires two external tools for running the tests during development.
./hack/ensure-kustomize.sh
./hack/tools/install_kubebuilder.sh
Both of the Tilt setups below will get you started developing CAPM3 in a local kind cluster. The main difference is the number of components you will build from source and the scope of the changes you'd like to make. If you only want to make changes in CAPM3, then follow CAPM3 instructions. This will save you from having to build all of the images for CAPI, which can take a while. If the scope of your development will span both CAPM3 and CAPI, then follow the CAPI and CAPM3 instructions.
In order to develop CAPM3 using Tilt, there is a requirement to have kind and Tilt installed.
hack/ensure-kind.sh
curl -fsSL https://raw.githubusercontent.com/tilt-dev/tilt/master/scripts/install.sh | bash
Once installed, you will need a tilt-settings.json. You can generate one using
make tilt-settings
The tilt-settings file can be customized. Tilt will also consider everything
under tilt.d
directory.
Then start Tilt:
make tilt-up
The Tilt dashboard can be accessed at http://localhost:10350. If tilt is running remotely, it can be accessed using ssh tunneling. An example of ~/.ssh/config
is show below.
Host tiltvm
HostName <SERVER_IP>
LocalForward 10350 127.0.0.1:10350
User ubuntu
Changes in the go code will trigger an update of the images running to run the latest code of CAPM3.
Tilt can be used in Metal3-dev-env, as long as the 03_launch_mgmt_cluster.sh script is NOT run. In that case Ironic needs to be deployed locally first, following the local Ironic setup
By default, the Cluster API components deployed by Tilt have experimental
features turned off. If you would like to enable these features, add
extra_args
as specified in
The Cluster API Book.
Once your kind management cluster is up and running, you can deploy an example cluster.
You can also deploy a flavor cluster as a local tilt resource.
To tear down the kind cluster built by the command above, just run:
make kind-reset
If you want to develop in both CAPI and CAPM3 at the same time, then this is the path for you.
First, you will need a kind setup with a registry. You can use the following command
make kind-create
To use Tilt for a simplified development workflow, follow
the instructions in the
cluster API repo. The instructions will walk you through cloning the Cluster
API (CAPI) repository and configuring Tilt to use kind
to deploy the cluster
api management components.
you may wish to checkout out the correct version of CAPI to match the version used in CAPM3
Note that tilt up
will be run from the cluster-api repository
directory and
the tilt-settings.json
file will point back to the
cluster-api-provider-metal3
repository directory. Any changes you make to the
source code in cluster-api
or cluster-api-provider-metal3
repositories will
automatically redeployed to the kind
cluster.
After you have cloned both repositories, your folder structure should look like:
|-- src/cluster-api-provider-metal3
|-- src/cluster-api (run `tilt up` here)
Checkout a stable CAPI release, for example:
git checkout release-0.3
Install kustomize from within the cluster-api repo
make kustomize
# verify installation
./hack/tools/bin/kustomize version
If the above installation does not work, install kustomize using a package manager and move the binary to hack/tools/bin/kustomize
in the cluster-api repo.
After configuring the environment variables, run the following to generate your
tilt-settings.json
file in the cluster API repository:
cat <<EOF > tilt-settings.json
{
"allowed_contexts": ["kind-capm3"],
"default_registry": "gcr.io/cluster-api-provider",
"deploy_cert_manager": true,
"preload_images_for_kind": true,
"kind_cluster_name": "capm3",
"provider_repos": ["../cluster-api-provider-metal3"],
"enable_providers": ["metal3", "docker", "kubeadm-bootstrap", "kubeadm-control-plane"],
"kustomize_substitutions": {
"DEPLOY_KERNEL_URL": "${DEPLOY_KERNEL_URL}",
"DEPLOY_RAMDISK_URL": "${DEPLOY_RAMDISK_URL}",
"IRONIC_INSPECTOR_URL": "${IRONIC_INSPECTOR_URL}",
"IRONIC_URL": "${IRONIC_URL}"
}
}
EOF
This requires the following environment variables to be exported :
DEPLOY_KERNEL_URL
DEPLOY_RAMDISK_URL
IRONIC_INSPECTOR_URL
IRONIC_URL
Those need to be set up accordingly with the local Ironic setup
The cluster-api management components that are deployed are configured at the
/config
folder of each repository respectively. Making changes to those files
will trigger a redeploy of the management cluster components.
If you want to develop on Baremetal Operator or IP Address Manager repository you can include them. First you need to clone the repositories you intend to modify locally, then modify your tilt-settings.json to point to the correct repositories. In case you have a tree like:
#-|
# |- cluster-api
# |- cluster-api-provider-metal3
# |- baremetal-operator
# |- ip-address-manager
and you create your tilt-settings.json
in the ./cluster-api or in the
./cluster-api-provider-metal3 folder. Then your tilt-settings.json file would
be :
{
"provider_repos": [ ... , "../baremetal-operator", "../ip-address-manager"],
"enable_providers": [ ... , "metal3-bmo", "metal3-ipam"],
}
The provider name for Baremetal Operator is metal3-bmo
and for IP Address
Manager metal3-ipam
. Those names are defined in their respective repository.
In case you are modifying the manifests for BMO or IPAM, then you should also modify the
kustomization.yaml
files in./config/bmo
and./config/ipam
to refer to the modified manifests locally instead of downloading released ones.
For a successful deployment, you need to have exported the proper variables. You can use the following command:
source ./examples/clusterctl-templates/example_variables.rc
However, the example variables do not guarantee a successful deployment, they need to be adapted. If deploying on Metal3-dev-env, please rather use the Metal3-dev-env deployment scripts that are tailored for it.
run tilt up ${flavors}
to spin up worker clusters represented by
a Tilt local resource. You can also modify flavors while Tilt is up by running
tilt args ${flavors}
Add your desired flavors to tilt_config.json:
{
"worker-flavors": ["default"]
}
Please note your tilt-settings.json must contain at minimum the following fields when using tilt resources to deploy cluster flavors:
{
"kustomize_substitutions": {
"DEPLOY_KERNEL_URL": "...",
"DEPLOY_RAMDISK_URL": "...",
"IRONIC_INSPECTOR_URL": "...",
"IRONIC_URL": "..."
}
}
If you wish to override the default variables for flavor workers, you can specify them as part of your tilt-settings.json as seen in the example below. Please note, the precedence of variables is as follows:
- explicitly defined vars for each flavor i.e.
worker-templates.flavors[0].WORKER_MACHINE_COUNT
- vars defined at 'metadata' level-- spans workers i.e.
metadata.WORKER_MACHINE_COUNT
- programmatically defined default vars i.e. everything except Ironic related URL variables "DEPLOY_KERNEL_URL", "DEPLOY_RAMDISK_URL", "IRONIC_INSPECTOR_URL" and "IRONIC_URL"
{
"kustomize_substitutions": {
"DEPLOY_KERNEL_URL": "...",
"DEPLOY_RAMDISK_URL": "...",
"IRONIC_INSPECTOR_URL": "...",
"IRONIC_URL": "..."
},
"worker-templates": {
"flavors": {
"default": {
"CLUSTER_NAME": "example-default-cluster-name",
}
},
"metadata": {,
"CONTROL_PLANE_MACHINE_COUNT": "1",
"KUBERNETES_VERSION": "v1.25.2",
"WORKER_MACHINE_COUNT": "2",
}
}
}
See the Kind docs for instructions on launching a Kind cluster and the Minikube docs for instructions on launching a Minikube cluster.
The following command will deploy the controllers from CAPI, CABPK and CAPM3 and
the requested CRDs and creates BareMetalHosts custom resources. The provider
uses the BareMetalHost
custom resource defined by the baremetal-operator
.
make deploy
When a Metal3Machine
is created, the provider looks for an available
BareMetalHost
object to claim and then sets it to be provisioned to fulfill the
request expressed by the Metal3Machine
. There’s no requirement to actually run
the baremetal-operator
to test the reconciliation logic of the provider.
Refer to the baremetal-operator developer documentation for instructions and tools for creating BareMetalHost objects.
You will first need to scale down the controller deployment:
kubectl scale -n capm3-system deployment.v1.apps/capm3-controller-manager --replicas 0
You can manually run the controller from outside of the cluster for development
and testing purposes. There’s a Makefile
target which makes this easy.
make run
You can follow the output on the console to see information about what the
controller is doing. You can also proceed to create/update/delete
Metal3Machines
and BareMetalHosts
to test the controller logic.
Make sure you run make deploy
and wait until all pods are in running
state
before deploying an example cluster with:
make deploy-examples
make delete-examples