All notable changes to this project will be documented in this file.
The format is based on Keep a Changelog, and this project adheres to Semantic Versioning.
0.68.0 - 2024-12-11
- Chart: Update
cluster
to v1.7.0.- Add
teleport-init
systemd unit to handle initial token setup beforeteleport
service starts - Improve
teleport
service reliability by adding proper file and service dependencies and pre-start checks
- Add
0.67.0 - 2024-11-28
Caution
The interface for setting annotations on the vsphereCluster
CR has changed.
Upgrade guide: how to migrate values (from v0.66.0)
The additionalVsphereClusterAnnotations
field is now an array of strings to accomodate for escape characters. Convert your dictionary to an array of strings as per the following example.
Old interface for annotations:
global:
providerSpecific:
additionalVsphereClusterAnnotations:
my-key: "my-value"
New interface for annotations:
global:
providerSpecific:
additionalVsphereClusterAnnotations:
- "my-key: value"
- Change
global.providerSpecific.additionalVsphereClusterAnnotations
attribute into an array.
0.66.0 - 2024-11-14
- Use Renovate to update
kube-vip
static pod manifest. - Updated
giantswarm/cluster
tov1.6.0
. - Update
kubectl
image used by IPAM job to1.29.9
. - Use init-container to prepare
/etc/hosts
file forkube-vip
.
0.65.2 - 2024-10-28
- Fix
kube-vip
static pod manifest for Kubernetes1.29
onwards.
0.65.1 - 2024-10-23
- Render Flatcar and Kubernetes version from cluster chart.
0.65.0 - 2024-10-15
- Support for Release CR's.
Migration steps
- In ConfigMap
<cluster name>-userconfig
set.Values.global.release.version
to the release version, e.g.27.0.0
. - In App
<cluster name>
set theversion
to an empty string.
- Update
kube-vip
static pod to v0.8.3. - Allow
.Values.global.managementCluster
in values schema.
0.64.0 - 2024-09-24
- Migrated all worker resources (
KubeadmConfigTemplate
,MachineDeployment
andMachineHealthCheck
) to be rendered from the sharedcluster
chart. - Render cleanup hook job using cluster chart.
0.63.0 - 2024-09-03
- Adding
global.connectivity.network.loadBalancers.numberOfIps
to specify the number of preassigned ips for load balancers. (New default is 3)
0.62.0 - 2024-08-29
- Allow adding custom annotations to the infrastructure cluster resource using
providerSpecific.additionalVsphereClusterAnnotations
value.
0.61.0 - 2024-08-23
Warning
This release adds all default apps to cluster-vsphere, so default-apps-vsphere App is not used anymore. Changes in cluster-vsphere are breaking and cluster upgrade requires manual steps where default-apps-vsphere App is removed before upgrading cluster-vsphere. See details below.
- Render capi-node-labeler App CR from cluster chart.
- Render cert-exporter App CR from cluster chart and add vSphere-specific cert-exporter config.
- Render cert-manager App CR from cluster chart and add vSphere-specific cert-manager config.
- Render chart-operator-extensions App CR from cluster chart.
- Render cilium HelmRelease CR from cluster chart and add vSphere-specific cilium config.
- Render cilium-servicemonitors App CR from cluster chart.
- Render coredns HelmRelease CR from cluster chart.
- Render etc-kubernetes-resources-count-exporter App CR from cluster chart.
- Render k8s-dns-node-cache App CR from cluster chart.
- Render metrics-server App CR from cluster chart.
- Render net-exporter App CR from cluster chart.
- Render network-policies HelmRelease CR from cluster chart and add vSphere-specific network-policies config.
- Render node-exporter App CR from cluster chart and add vSphere-specific node-exporter config.
- Render observability-bundle App CR from cluster chart.
- Render observability-policies App CR from cluster chart.
- Render security-bundle App CR from cluster chart.
- Render teleport-kube-agent App CR from cluster chart.
- Render vertical-pod-autoscaler App CR from cluster chart.
- Render vertical-pod-autoscaler-crd HelmRelease CR from cluster chart.
- Render HelmRepository CRs from cluster chart.
- Add missing Helm value .Values.global.controlPlane.apiServerPort.
- Add Makefile
template
target that renders manifests with CI values from the chart. - Add Makefile
generate
target that normalizes and validates schema, generates docs and Helm values, and updates Helm dependencies.
- Remove cilium HelmRelease.
- Remove coredns HelmRelease.
- Remove network-policies HelmRelease.
- Remove HelmRepository CRs.
The steps to upgrade a workload cluster, with unifying cluster-vsphere and default-apps-vsphere, are the following:
- Upgrade default-apps-vsphere App to the v0.16.0 release.
- Update default-apps-vsphere Helm value
.Values.deleteOptions.moveAppsHelmOwnershipToClusterVSphere
totrue
.- All App CRs, except observability-bundle and security-bundle, will get
app-operator.giantswarm.io/paused: true
annotation, so wait few minutes for Helm post-upgrade hook to apply the change to all required App CRs.
- All App CRs, except observability-bundle and security-bundle, will get
- Delete default-apps-vsphere CR.
⚠️ In case you are removing default-apps-vsphere App CR from your gitops repo which is using Flux, and depending on how Flux is configured, default-apps-vsphere App CR may or may not get deleted from the management cluster. In case Flux does not delete default-apps-vsphere App CR from the management cluster, make sure to delete it manually.- If default-apps-vsphere have specific configure in the values, transfer these values to the cluster app values.
- App CRs (on the MC) for all default apps will get deleted. Wait few minutes for this to happen.
- Chart CRs on the workload cluster will remain untouched, so all apps will continue running.
- Upgrade cluster-vsphere App CR to the v0.61.0 release.
- cluster-vsphere will deploy all default apps, so wait a few minutes for all Apps to be successfully deployed.
- Chart resources on the workload cluster will get updated, as newly deployed App resources will take over the reconciliation of the existing Chart resources.
We're almost there, with just one more issue to fix manually.
VPA CRD used to installed as an App resource from default-apps-vsphere, and now it's being installed as a HelmRelease from cluster-vsphere. Now, as a consequence of the above upgrade, we have the following situation:
- default-apps-vsphere App has been deleted, but the vertical-pod-autoscaler-crd Chart CRs remained in the workload cluster.
- cluster-vsphere has been upgraded, so now it also installs vertical-pod-autoscaler-crd HelmRelease.
- outcome: we now have vertical-pod-autoscaler-crd HelmRelease in the MC and vertical-pod-autoscaler-crd Chart CR in the WC.
Now we will remove the leftover vertical-pod-autoscaler-crd Chart CR in a safe way:
- Pause vertical-pod-autoscaler-crd Chart CR.
Add annotation chart-operator.giantswarm.io/paused: "true"
to vertical-pod-autoscaler-crd Chart CR in the workload cluster:
kubectl annotate -n giantswarm chart vertical-pod-autoscaler-crd chart-operator.giantswarm.io/paused="true" --overwrite
- Delete vertical-pod-autoscaler-crd Chart CR in the workload cluster.
kubectl delete -n giantswarm chart vertical-pod-autoscaler-crd
The command line will probably hang, as the chart-operator finalizer has is not getting removed (vertical-pod-autoscaler-crd Chart CR has been paused). Proceed to the next step to remove the finalizer and unblock the deletion.
- Remove finalizers from the vertical-pod-autoscaler-crd Chart CR
Open another terminal window and run the following command to remove the vertical-pod-autoscaler-crd Chart CR finalizers:
kubectl patch chart vertical-pod-autoscaler-crd -n giantswarm --type=json -p='[{"op": "remove", "path": "/metadata/finalizers"}]'
This will unblock the deletion and vertical-pod-autoscaler-crd will get removed, without actually deleting VPA CustomResourceDefinition.
From now on, VPA CustomResourceDefinition will be maintained by the vertical-pod-autoscaler HelmRelease on the management cluster.
0.60.1 - 2024-08-23
- Rename caFile to caPem in values schema.
0.60.0 - 2024-08-22
Caution
It is important that you check each of the sections in the upgrade guide below. Note that some may not apply to your specific cluster configuration. However, the cleanup section must always be run against the cluster values.
Upgrade guide: how to migrate values (from v0.59.0)
Use the snippets below if the section applies to your chart's values:
If the controlPlane endpoint IP (loadbalancer for the Kubernetes API) has been statically assigned (this likely will not apply to workload clusters) then this value will need to be duplicated to the extraCertificateSANs list. Also, any additional certificate SANs must be added to the extraCertificateSANs list.
yq eval --inplace 'with(select(.global.connectivity.network.controlPlaneEndpoint.host != null); .cluster.internal.advancedConfiguration.controlPlane.apiServer.extraCertificateSANs += [ .global.connectivity.network.controlPlaneEndpoint.host ]) |
with(select(.internal.apiServer.certSANs != null); .cluster.internal.advancedConfiguration.controlPlane.apiServer.extraCertificateSANs += [ .internal.apiServer.certSANs[] ])' values.yaml
The default list is here. If you have not extended this list then you do not need to provide a list of admission plugins at all (defaults will be used from the cluster chart). If this is the case, please ignore the following command.
yq eval --inplace 'with(select(.internal.apiServer.enableAdmissionPlugins != null); .cluster.providerIntegration.controlPlane.kubeadmConfig.clusterConfiguration.apiServer.additionalAdmissionPlugins = .internal.apiServer.enableAdmissionPlugins)' values.yaml
There is no default list of feature gates in the shared cluster chart, so if you have any values under .internal.apiServer.featureGates
then these must be migrated to the new location.
yq eval --inplace 'with(select(.internal.apiServer.featureGates != null); .cluster.providerIntegration.controlPlane.kubeadmConfig.clusterConfiguration.apiServer.featureGates = .internal.apiServer.featureGates)' values.yaml
caFile
has been renamed to caPem
.
yq eval --inplace 'with(select(.global.controlPlane.oidc.caFile != null); .global.controlPlane.oidc.caPem = .global.controlPlane.oidc.caFile)' values.yaml
If you are providing additional trusted CA keys for SSH authentication (other than the default Giant Swarm key) then these need to migrated to the new location.
yq eval --inplace 'with(select(.global.connectivity.shell.sshTrustedUserCAKeys != null); .cluster.providerIntegration.connectivity.sshSsoPublicKey = .global.connectivity.shell.sshTrustedUserCAKeys)' values.yaml
If your cluster is behind an upstream proxy (if .global.connectivity.proxy.enabled: true
) then the proxy configuration must also be added to the cluster chart's values.
httpProxy
: upstream proxy protocol, address and port (e.g.http://proxy-address:port
)httpsProxy
: upstream proxy protocol, address and port (e.g.http://proxy-address:port
)noProxy
: comma-separated list of domains and IP CIDRs which should not be proxied (e.g.10.10.10.0/24,internal.domain.com
)
Additional notes:
- Encryption is always enabled with the shared cluster chart, so this toggle is removed entirely (
.internal.enableEncryptionProvider
). - OIDC
groupsPrefix
andusernamePrefix
are removed. - Upstream proxy configuration is no longer read from the
.global.connectivity.proxy.secretName
value.
Final tidyup to remove deprecated values:
yq eval --inplace 'del(.internal.apiServer.enableAdmissionPlugins) |
del(.internal.apiServer.featureGates) |
del(.internal.apiServer.certSANs) |
del(.internal.enableEncryptionProvider) |
del(.global.controlPlane.oidc.caFile) |
del(.global.controlPlane.oidc.groupsPrefix) |
del(.global.controlPlane.oidc.usernamePrefix) |
del(.global.connectivity.shell.sshTrustedUserCAKeys) |
del(.global.connectivity.proxy.secretName) |
del(.internal.apiServer) |
del(.internal.controllerManager) |
del(.internal.enableEncryptionProvider) ' values.yaml
- Use
giantswarm/cluster
chart to renderKubeadmControlPlane
resource. - Update giantswarm/cluster chart to 1.2.1.
0.59.0 - 2024-08-12
- Set
kubeProxyReplacement
to'true'
instead of deprecated valuestrict
in cilium values. - Make default storage class configurable.
0.58.3 - 2024-08-04
- Allow additional properties on machine templates to offer wider CAPV configurations.
0.58.2 - 2024-08-01
- Disable IPAM for service load balancers if
.global.connectivity.network.loadBalancers.cidrBlocks
is set.
0.58.1 - 2024-07-31
- Change
cluster
in kubectl patch command tocluster.cluster.x-k8s.io
inside hooks in case more than one API group is present.
0.58.0 - 2024-07-30
- Fix syntax error with helmrelease cleanup hook.
- Update
ipam
API versions tov1beta1
.
0.57.1 - 2024-07-24
0.57.0 - 2024-07-22
Caution
The cluster name must be added to the cluster values when upgrading to this chart release.
How to migrate values (from v0.56.1)
The cluster's name must be added to the cluster values in order to satisfy the updated values
schema. This can be done by adding the existing cluster name to the cluster
values. For example,
where the cluster is named test
:
global:
metadata:
name: test
- Initial integration of shared
cluster
chart to renderCluster
resource. - Corrected cluster names in CI values.
0.56.1 - 2024-07-16
- Unpause Cluster resource as part of cleanup hook after deletion in order to prevent leftover resources.
0.56.0 - 2024-07-10
Caution
Upgrading to this chart release will cause all worker nodes to be replaced.
How to migrate values (from `v0.54.0` or later)
Using yq
, migrate to the new values layout with the following command:
yq eval --inplace 'with(select(.global.nodeClasses != null); .global.nodeClasses as $classes | with(.global.nodePools[]; . *= $classes[.class])) |
del(.global.nodePools[].class) |
del(.global.nodeClasses)' values.yaml
- Move Helm values from each
.global.nodeClasses.$<class>
to any nodePool which references that class. - Deleted Helm values property
.global.nodeClasses
.
- Correct default values for
network
s in values schema.
0.55.0 - 2024-06-25
- Bump
cloud-provider-vsphere
to1.7.0
for Kubernetes 1.27 compatibility.- Update vSphere CSI to
3.2.0
. - Update vSphere CPI to
1.27.0
. - Update kube-vip to
0.8.0
. - Update kube-vip-cloud-provider to
0.0.5
.
- Update vSphere CSI to
0.54.0 - 2024-06-25
- Add
.global.connectivity.localRegistryCache
Helm values and support for in-cluster, local registry cache mirrors in containerd configuration. In such cases, the registry should be exposed via node ports and containerd connects via that port at 127.0.0.1 via HTTP (only allowed for this single use case).
- Fixed
containerd
config file generation when multiple registries are set with authentication
- Stop setting
defaultPolicies.enabled=true
incilium-app
wheninternal.ciliumNetworkPolicy.enabled=true
after all clusters are migrated. - Stop setting
extraPolicies.remove=true
incilium-app
after all clusters are migrated.
0.53.1 - 2024-06-09
- Refer the API group (
ipam.cluster.x-k8s.io
) ofcluster-api-ipam-provider-in-cluster
foripaddresses
CRs to not use the built-in Kubernetes group (networking.k8s.io/v1alpha1
).
0.53.0 - 2024-06-06
- Remove kube-vip values to rely on the defaults of
cloud-provider-vsphere-app
. - Bump k8s version from
1.26.15
to1.27.14
.
0.52.0 - 2024-05-23
- Normalise JSON schema.
- Remove unused values from schema.
- Update example manifests post-refactor.
- Improve values schema with definitions to make it more DRY.
- Bump k8s version from
1.25.16
to1.26.15
.
0.51.0 - 2024-05-16
How to migrate values
Using yq
, migrate to the new values layout with the following command:
#!/bin/bash
yq eval --inplace 'with(select(.metadata != null); .global.metadata = .metadata) |
with(select(.clusterDescription != null); .global.metadata.description = .clusterDescription) |
with(select(.organization != null); .global.metadata.organization = .organization) |
with(select(.clusterLabels != null); .global.metadata.labels = .clusterLabels) |
with(select(.servicePriority != null); .global.metadata.servicePriority = .servicePriority) |
with(select(.connectivity != null); .global.connectivity = .connectivity) |
with(select(.osUsers != null); .global.connectivity.shell.osUsers = .osUsers) |
with(select(.sshTrustedUserCAKeys != null); .global.connectivity.shell.sshTrustedUserCAKeys = .sshTrustedUserCAKeys) |
with(select(.proxy != null); .global.connectivity.proxy = .proxy) |
with(select(.baseDomain != null); .global.connectivity.baseDomain = .baseDomain) |
with(select(.controlPlane != null); .global.controlPlane = .controlPlane) |
with(select(.oidc != null); .global.controlPlane.oidc = .oidc) |
with(select(.nodePools != null); .global.nodePools = .nodePools) |
with(select(.vcenter != null); .global.providerSpecific.vcenter = .vcenter) |
with(select(.cluster.kubernetesVersion != null); .internal.kubernetesVersion = .cluster.kubernetesVersion) |
with(select(.cluster.enableEncryptionProvider != null); .internal.enableEncryptionProvider = .cluster.enableEncryptionProvider) |
with(select(.controllerManager.featureGates != null); .internal.controllerManager.featureGates = (.controllerManager.featureGates | split(","))) |
with(select(.apiServer.enableAdmissionPlugins != null); .internal.apiServer.enableAdmissionPlugins = (.apiServer.enableAdmissionPlugins | split(","))) |
with(select(.apiServer.featureGates != null); .internal.apiServer.featureGates = (.apiServer.featureGates | split(","))) |
with(select(.apiServer.certSANs != null); .internal.apiServer.certSANs = .apiServer.certSANs) |
with(select(.kubectlImage != null); .internal.kubectlImage = .kubectlImage) |
with(select(.nodeClasses != null); .global.nodeClasses = .nodeClasses) |
del(.metadata) |
del(.clusterDescription) |
del(.organization) |
del(.clusterLabels) |
del(.servicePriority) |
del(.connectivity) |
del(.osUsers) |
del(.sshTrustedUserCAKeys) |
del(.proxy) |
del(.baseDomain) |
del(.controlPlane) |
del(.oidc) |
del(.nodePools) |
del(.vcenter) |
del(.cluster) |
del(.controllerManager) |
del(.apiServer) |
del(.kubectlImage) |
del(.nodeClasses)' values.yaml
- Move Helm values property
.Values.metadata
to.Values.global.metadata
. - Move Helm values property
.Values.clusterDescription
to.Values.global.metadata.description
. - Move Helm values property
.Values.organization
to.Values.global.metadata.organization
. - Move Helm values property
.Values.clusterLabels
to.Values.global.metadata.labels
. - Move Helm values property
.Values.servicePriority
to.Values.global.metadata.servicePriority
. - Move Helm values property
.Values.connectivity
to.Values.global.connectivity
. - Move Helm values property
.Values.proxy
to.Values.global.connectivity.proxy
. - Move Helm values property
.Values.osUsers
to.Values.global.connectivity.shell.osUsers
. - Move Helm values property
.Values.sshTrustedUserCAKeys
to.Values.global.connectivity.shell.sshTrustedUserCAKeys
. - Move Helm values property
.Values.baseDomain
to.Values.global.connectivity.baseDomain
. - Move Helm values property
.Values.controlPlane
to.Values.global.controlPlane
. - Move Helm values property
.Values.oidc
to.Values.global.controlPlane.oidc
. - Move Helm values property
.Values.nodePools
to.Values.global.nodePools
. - Move Helm values property
.Values.vcenter
to.Values.global.providerSpecific.vcenter
. - Move Helm values property
.Values.controllerManager.featureGates
to.Values.internal.controllerManager.featureGates
and convert from string to array. - Move Helm values property
.Values.apiServer.enableAdmissionPlugins
to.Values.internal.apiServer.enableAdmissionPlugins
and convert from string to array. - Move Helm values property
.Values.apiServer.featureGates
to.Values.internal.apiServer.featureGates
and convert from string to array. - Move Helm values property
.Values.apiServer.certSANs
to.Values.internal.apiServer.certSANs
. - Move Helm values property
.Values.kubectlImage
to.Values.internal.kubectlImage
. - Move Helm values property
.Values.nodeClasses
to.Values.global.nodeClasses
.
0.50.0 - 2024-04-23
- No major change in
v0.50.0
, except that we are moving to a release based upgrade cycle with Kubernetes version, VM template and other defaults are set in the chart values. They shouldn't be overridden as they are managed by Giant Swarm. - Bump
kube-vip
tov0.8.0
.
0.10.3 - 2024-04-08
- Move extraPolicies from cilium-app to network-policies-app.
- Add
svc-lb-ips
as default IP Pool for service of type load balancer in workload clusters.
0.10.2 - 2024-04-02
- Pull
kube-vip
image from Azure CR. - Update teleport node labels - add
ins=
label and removecluster=
label condition check, such that MC nodes have this label.
0.10.1 - 2024-03-07
- Bump Cilium to
0.21.0
.
0.10.0 - 2024-02-27
- Add CiliumNetworkPolicies for the cleanup job.
- Add flags to disable PSPs.
- Add
global.metadata.preventDeletion
to add the deletion prevention label to Cluster resources.
- Switch container registry to
gsoci.azurecr.io
. - Adapt cleanup hook for cluster policies.
0.9.9 - 2024-02-14
- Bump netpol app which disables the coredns
CiliumClusterwideNetworkPolicy
.
0.9.8 - 2024-02-12
- Include support for schemadocs to generate Chart README file
0.9.7 - 2024-01-23
- Add teleport.service: Secure SSH access via Teleport
- Bump Cilium to
0.19.2
.
0.9.6 - 2023-12-14
- Add vSphere icon.
- Fix cleanup of netpol helmrelease.
0.9.5 - 2023-12-12
- Fix templating issues of feature-gates.
0.9.4 - 2023-12-11
0.9.3 - 2023-12-07
- If
connectivity.network.allowAllEgress
is false (which is the default value), the deny-all network policies for namespacesgiantswarm
andkube-system
will be applied to the resulting cluster. In terms of API, this is a compatible change but in terms of internal behavior it can be potentially 💥 Breaking:.
- Remove
TTLAfterFinished
flag for Kubernetes 1.25 compatibility (enabled by default). - Remove
ExpandPersistentVolumes
flag for Kubernetes 1.27 compatibility (enabled by default). - Remove
logtostderr
for Kubernetes 1.27 compatibility (output is logged to stderr by default).
0.9.2 - 2023-11-15
- Minor fix and use of
--ignore-not-found
in IPAM.
0.9.1 - 2023-11-15
- Bump
cilium
to0.17.0
. - Bump
coredns
to1.19.0
. - Enable renovate for
cilium
andcoredns
.
0.9.0 - 2023-11-08
- 💥 Breaking: Switch from Ubuntu to Flatcar and start support ignition.
0.8.0 - 2023-10-31
- Add IPv4 addresses management (ipam) for WC's
kube-vip-cloud-provider
. If the ipPool is specified, currently one IP is requested from it and is added to the end of the list for this controller.kube-vip-cloud-provider
is part of our Service-lvl load balancer solution in WC and at least 1 public IP is always needed for the ingress controller to be able to expose its stuff.
0.7.1 - 2023-09-04
- Remove leftover
helmchart
CRs.
0.7.0 - 2023-08-17
0.6.2 - 2023-08-01
- Consolidate containerd
config.toml
into single file to address #1737 - Add host OS user
nobody
toroot
group to enable node-exporter'sfilesystem
collector to access the host filesystem. - Bump
cilium
version to0.10.0
0.6.1 - 2023-07-13
- Add a way to customize the timeout in
HelmRelease.spec
. - Set value for
controller-manager
terminated-pod-gc-threshold
to125
( consistent with vintage ) - Bump
cloud-provider-vsphere
version to1.5.0
0.6.0 - 2023-07-04
- Add IPv4 addresses management (ipam) for WC's control planes.
- Add
connectivity.network.controlPlaneEndpoint.host
tocertSANs
list. - Bind
kube-scheduler
metrics to 0.0.0.0.
0.5.1 - 2023-06-07
- Bump the version of coredns-app to
1.16.0
0.5.0 - 2023-05-22
- 💥 Breaking: Stop deploying default network policies with the
cilium-app
. This means the cluster will be more locked down and all network traffic is blocked by default. Can be disabled withnetwork.allowAllEgress
setting. Requiresdefault-apps-vsphere@v0.9.2
. - Bumped default k8s version to
1.24
, this might be 💥 Breaking: .cluster.kubernetesVersion
:v1.22.5+vmware.1
->v1.24.11
- default machine template
ubuntu-2004-kube-v1.22.5+vmware.1
->ubuntu-2004-kube-v1.24.11
.controlPlane.image.repository
:projects.registry.vmware.com/tkg
->registry.k8s.io
- Add audilog configuration.
- 💥 Breaking: Refactor api for network parameters to apply the standard interface for all providers.
- Set
/var/lib/kubelet
permissions to0750
to fixnode-exporter
issue.
0.4.0 - 2023-04-24
- Add
MachineHealthCheck
for worker nodes. - Add
loadBalancersCidrBlocks
parameter that is used by kube-vip forLoadBalancer
services. - Add
apiServer.certSANs
option.
- Improve schema and ci values.
- 💥 Breaking: Change default pod network and service network to 10.244.0.0/16 and 10.96.0.0/16.
- 💥 Breaking: Install CoreDNS (
coredns-app
) usingHelmRelease
CR and stop deploying it withcluster-shared
resource set.
- Set
.network.servicesCidrBlocks
value in Cluster CR.
- Remove
cluster-shared
dependency.
0.3.1 - 2023-04-05
- Add
default-test
HelmRepository (catalog) for debugging.
- Bump
cloud-provider-vsphere
version to1.3.3
.
- Remove unnecessary labels from HelmRepository CR.
0.3.0 - 2023-03-27
- 💥 Breaking: Use cilium kube-proxy replacement.
- Bump
cloud-provider-vsphere
version to1.3.2
. - Use release name instead of
cluster.name
. - Move
organization
to root level for uniformity.
0.2.1 - 2023-03-21
- Add missing files for CoreDNS configuration by cluster-shared.
0.2.0 - 2023-03-20
- Allow setting etcd image repository and tag.
- Set the default etcd version to 3.5.4 (kubeadm default is 3.5.0 which is not recommended in production).
- Set the default etcd image to retagged Giant Swarm one.
0.1.2 - 2022-05-09
0.1.1 - 2022-03-29
- Add CicleCI configuration.
0.1.0 - 2022-03-29
- Initial chart implementation.