# Oracle Cloud Infrastructure This tutorial describes how to setup ExternalDNS for usage within a Kubernetes cluster using OCI DNS. Make sure to use the latest version of ExternalDNS for this tutorial. ## Creating an OCI DNS Zone Create a DNS zone which will contain the managed DNS records. Let's use `example.com` as a reference here. Make note of the OCID of the compartment in which you created the zone; you'll need to provide that later. For more information about OCI DNS see the documentation [here][1]. ## Using Private OCI DNS Zones By default, the ExternalDNS OCI provider is configured to use Global OCI DNS Zones. If you want to use Private OCI DNS Zones, add the following argument to the ExternalDNS controller: ``` --oci-zone-scope=PRIVATE ``` To use both Global and Private OCI DNS Zones, set the OCI Zone Scope to be empty: ``` --oci-zone-scope= ``` ## Deploy ExternalDNS Connect your `kubectl` client to the cluster you want to test ExternalDNS with. The OCI provider supports two authentication options: key-based and instance principals. ### Key-based We first need to create a config file containing the information needed to connect with the OCI API. Create a new file (oci.yaml) and modify the contents to match the example below. Be sure to adjust the values to match your own credentials, and the OCID of the compartment containing the zone: ```yaml auth: region: us-phoenix-1 tenancy: ocid1.tenancy.oc1... user: ocid1.user.oc1... key: | -----BEGIN RSA PRIVATE KEY----- -----END RSA PRIVATE KEY----- fingerprint: af:81:71:8e... # Omit if there is not a password for the key passphrase: Tx1jRk... compartment: ocid1.compartment.oc1... ``` Create a secret using the config file above: ```shell $ kubectl create secret generic external-dns-config --from-file=oci.yaml ``` ### OCI IAM Instance Principal If you're running ExternalDNS within OCI, you can use OCI IAM instance principals to authenticate with OCI. This obviates the need to create the secret with your credentials. You'll need to ensure an OCI IAM policy exists with a statement granting the `manage dns` permission on zones and records in the target compartment to the dynamic group covering your instance running ExternalDNS. E.g.: ``` Allow dynamic-group to manage dns in compartment id ``` You'll also need to add the `--oci-auth-instance-principal` flag to enable this type of authentication. Finally, you'll need to add the `--oci-compartment-ocid=ocid1.compartment.oc1...` flag to provide the OCID of the compartment containing the zone to be managed. For more information about OCI IAM instance principals, see the documentation [here][2]. For more information about OCI IAM policy details for the DNS service, see the documentation [here][3]. ### OCI IAM Workload Identity If you're running ExternalDNS within an OCI Container Engine for Kubernetes (OKE) cluster, you can use OCI IAM Workload Identity to authenticate with OCI. You'll need to ensure an OCI IAM policy exists with a statement granting the `manage dns` permission on zones and records in the target compartment covering your OKE cluster running ExternalDNS. E.g.: ``` Allow any-user to manage dns in compartment where all {request.principal.type='workload',request.principal.cluster_id='',request.principal.service_account='external-dns'} ``` You'll also need to create a new file (oci.yaml) and modify the contents to match the example below. Be sure to adjust the values to match your region and the OCID of the compartment containing the zone: ```yaml auth: region: us-phoenix-1 useWorkloadIdentity: true compartment: ocid1.compartment.oc1... ``` Create a secret using the config file above: ```shell $ kubectl create secret generic external-dns-config --from-file=oci.yaml ``` ## Manifest (for clusters with RBAC enabled) Apply the following manifest to deploy ExternalDNS. ```yaml apiVersion: v1 kind: ServiceAccount metadata: name: external-dns --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: name: external-dns rules: - apiGroups: [""] resources: ["services","endpoints","pods"] verbs: ["get","watch","list"] - apiGroups: ["extensions","networking.k8s.io"] resources: ["ingresses"] verbs: ["get","watch","list"] - apiGroups: [""] resources: ["nodes"] verbs: ["list"] --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRoleBinding metadata: name: external-dns-viewer roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: external-dns subjects: - kind: ServiceAccount name: external-dns namespace: default --- apiVersion: apps/v1 kind: Deployment metadata: name: external-dns spec: strategy: type: Recreate selector: matchLabels: app: external-dns template: metadata: labels: app: external-dns spec: serviceAccountName: external-dns containers: - name: external-dns image: registry.k8s.io/external-dns/external-dns:v0.15.0 args: - --source=service - --source=ingress - --provider=oci - --policy=upsert-only # prevent ExternalDNS from deleting any records, omit to enable full synchronization - --txt-owner-id=my-identifier # Specifies the OCI DNS Zone scope, defaults to GLOBAL. # May be GLOBAL, PRIVATE, or an empty value to specify both GLOBAL and PRIVATE OCI DNS Zones # - --oci-zone-scope=GLOBAL # Specifies the zone cache duration, defaults to 0s. If set to 0s, the zone cache is disabled. # Use of zone caching is recommended to reduce the amount of requests sent to OCI DNS. # - --oci-zones-cache-duration=0s volumeMounts: - name: config mountPath: /etc/kubernetes/ volumes: - name: config secret: secretName: external-dns-config ``` ## Verify ExternalDNS works (Service example) Create the following sample application to test that ExternalDNS works. > For services ExternalDNS will look for the annotation `external-dns.alpha.kubernetes.io/hostname` on the service and use the corresponding value. ```yaml apiVersion: v1 kind: Service metadata: name: nginx annotations: external-dns.alpha.kubernetes.io/hostname: example.com spec: type: LoadBalancer ports: - port: 80 name: http targetPort: 80 selector: app: nginx --- apiVersion: apps/v1 kind: Deployment metadata: name: nginx spec: selector: matchLabels: app: nginx template: metadata: labels: app: nginx spec: containers: - image: nginx name: nginx ports: - containerPort: 80 name: http ``` Apply the manifest above and wait roughly two minutes and check that a corresponding DNS record for your service was created. ``` $ kubectl apply -f nginx.yaml ``` [1]: https://docs.cloud.oracle.com/iaas/Content/DNS/Concepts/dnszonemanagement.htm [2]: https://docs.cloud.oracle.com/iaas/Content/Identity/Reference/dnspolicyreference.htm [3]: https://docs.cloud.oracle.com/iaas/Content/Identity/Tasks/callingservicesfrominstances.htm