Install Multi-Primary
Follow this guide to install the Istio control plane on both cluster1 and
cluster2, making each a primary cluster. Both
clusters reside on the network1 network, meaning there is direct
connectivity between the pods in both clusters.
Before proceeding, be sure to complete the steps under before you begin.
In this configuration, each control plane observes the API Servers in both clusters for endpoints.
Service workloads communicate directly (pod-to-pod) across cluster boundaries.
Configure cluster1 as a primary
Create the istioctl configuration for cluster1:
Install Istio as primary in cluster1 using istioctl and the IstioOperator API.
$ cat <<EOF > cluster1.yaml
apiVersion: install.istio.io/v1alpha1
kind: IstioOperator
spec:
values:
global:
meshID: mesh1
multiCluster:
clusterName: cluster1
network: network1
EOFApply the configuration to cluster1:
$ istioctl install --context="${CTX_CLUSTER1}" -f cluster1.yamlInstall Istio as primary in cluster1 using the following Helm commands:
Install the base chart in cluster1:
$ helm install istio-base istio/base -n istio-system --kube-context "${CTX_CLUSTER1}"Then, install the istiod chart in cluster1 with the following multi-cluster settings:
$ helm install istiod istio/istiod -n istio-system --kube-context "${CTX_CLUSTER1}" --set global.meshID=mesh1 --set global.multiCluster.clusterName=cluster1 --set global.network=network1Configure cluster2 as a primary
Create the istioctl configuration for cluster2:
Install Istio as primary in cluster2 using istioctl and the IstioOperator API.
$ cat <<EOF > cluster2.yaml
apiVersion: install.istio.io/v1alpha1
kind: IstioOperator
spec:
values:
global:
meshID: mesh1
multiCluster:
clusterName: cluster2
network: network1
EOFApply the configuration to cluster2:
$ istioctl install --context="${CTX_CLUSTER2}" -f cluster2.yamlInstall Istio as primary in cluster2 using the following Helm commands:
Install the base chart in cluster2:
$ helm install istio-base istio/base -n istio-system --kube-context "${CTX_CLUSTER2}"Then, install the istiod chart in cluster2 with the following multi-cluster settings:
$ helm install istiod istio/istiod -n istio-system --kube-context "${CTX_CLUSTER2}" --set global.meshID=mesh1 --set global.multiCluster.clusterName=cluster2 --set global.network=network1Enable Endpoint Discovery
Install a remote secret in cluster2 that provides access to cluster1’s API server.
$ istioctl create-remote-secret \
--context="${CTX_CLUSTER1}" \
--name=cluster1 | \
kubectl apply -f - --context="${CTX_CLUSTER2}"Install a remote secret in cluster1 that provides access to cluster2’s API server.
$ istioctl create-remote-secret \
--context="${CTX_CLUSTER2}" \
--name=cluster2 | \
kubectl apply -f - --context="${CTX_CLUSTER1}"Congratulations! You successfully installed an Istio mesh across multiple primary clusters!
Next Steps
You can now verify the installation.
Cleanup
Uninstall Istio from both cluster1 and cluster2 using the same mechanism you installed Istio with (istioctl or Helm).
Uninstall Istio in cluster1:
$ istioctl uninstall --context="${CTX_CLUSTER1}" -y --purge
$ kubectl delete ns istio-system --context="${CTX_CLUSTER1}"Uninstall Istio in cluster2:
$ istioctl uninstall --context="${CTX_CLUSTER2}" -y --purge
$ kubectl delete ns istio-system --context="${CTX_CLUSTER2}"Delete Istio Helm installation from cluster1:
$ helm delete istiod -n istio-system --kube-context "${CTX_CLUSTER1}"
$ helm delete istio-base -n istio-system --kube-context "${CTX_CLUSTER1}"Delete the istio-system namespace from cluster1:
$ kubectl delete ns istio-system --context="${CTX_CLUSTER1}"Delete Istio Helm installation from cluster2:
$ helm delete istiod -n istio-system --kube-context "${CTX_CLUSTER2}"
$ helm delete istio-base -n istio-system --kube-context "${CTX_CLUSTER2}"Delete the istio-system namespace from cluster2:
$ kubectl delete ns istio-system --context="${CTX_CLUSTER2}"(Optional) Delete CRDs installed by Istio:
Deleting CRDs permanently removes any Istio resources you have created in your clusters. Delete Istio CRDs installed in your clusters by running:
$ kubectl get crd -oname --context "${CTX_CLUSTER1}" | grep --color=never 'istio.io' | xargs kubectl delete --context "${CTX_CLUSTER1}"
$ kubectl get crd -oname --context "${CTX_CLUSTER2}" | grep --color=never 'istio.io' | xargs kubectl delete --context "${CTX_CLUSTER2}"