Deploy Data Source Connector
To register your data sources with the IBM Cloud Backup and Recovery service, you need to establish connectivity between your source and the service by using a Data Source Connection. A Data Source Connection consists of one or more Data Source Connectors that facilitate the movement of data between your data sources and the IBM Cloud Backup and Recovery service.
The deployment method for Data Source Connectors varies based on your source type:
-
For VM-based sources (Physical Servers, Microsoft SQL Server, Oracle Server): Deploy the Data Source Connector as a virtual machine (VM) using an installer OVA in your VMware environment, on a vCenter or ESXi host that has access to your data sources and meets the system and firewall requirements.
-
For Kubernetes/OpenShift clusters: Deploy the Data Source Connector using Helm charts directly on your cluster. See Install Data Source Connector for Kubernetes/OpenShift for detailed instructions.
Alternatively, you can also refer to the terraform IBM Backup & Recovery for IKS/ROKS with Data Source Connector module that offers ready-to-use code and examples for integrating the Data Source connector.
Data Source Connector Requirements
Before deploying the Data Source Connector, review and understand the following requirements that are needed for the VM(s) that you need to provision:
Supported Sources
You can deploy Data Source Connectors for the following sources:
- Physical Server
- Microsoft SQL Server
- Oracle Server
- Kubernetes/OpenShift
Data Source Connector System Prerequisites
Make sure that the Data Source Connector VM that you deploy for your Data Source Connection meets the following system requirements:
- 4 CPUs
- 16 GB RAM
- 171 GB disk space
- Outbound Internet connection
Port Requirements
Ensure that the following ports are open to allow communication between the Data Source Connector(s) and the data sources.
Physical Servers
| Source | Destination | Port | Protocol | Purpose |
|---|---|---|---|---|
| Data Source Connector | Physical Windows or Linux Server | 50051 | TCP | Required for Backup and Recovery operations. |
| Local Host (Physical Windows or Linux Server) | Local Host (Physical Windows or Linux Server) | 59999 | TCP | Required for local-to-local communication for self-monitoring and debugging purposes. |
Microsoft SQL Servers
| Source | Destination | Port | Protocol | Purpose |
|---|---|---|---|---|
| Data Source Connector | MS SQL Host | 50051 | TCP | Required for Backup and Recovery operations. |
| MS SQL Host | Data Source Connector | 11113,11117 | TCP | Required for Backup and Recovery operations. |
| MS SQL Host | Agent running on the MS SQL Host | 1433 | TCP | Default TCP port for MS SQL instances. Ensure that the port is open to allow communication between the MS SQL instance and the Agent. |
Oracle Servers
Ensure that the following ports are open to allow communication between the Data Source Connector(s) and Oracle Server:
| Source | Destination | Port | Protocol | Purpose |
|---|---|---|---|---|
| Data Source Connector | Oracle Server | 50051 | TCP | Required for Backup and Recovery operations. |
| Oracle Server | Data Source Connector | 111, 2049 | TCP | Required for Backup and Recovery operations in Linux servers. |
| Oracle Server | Data Source Connector | 11113, 11117 | TCP | Required for Backup and Recovery operations in Windows servers. |
| Local Host (Physical Windows or Linux Server) | Local Host (Physical Windows or Linux Server) | 59999 | TCP | Required for local-to-local communication for self-monitoring and debugging purposes. |
IBM Backup Service
Ensure that the following ports are open to allow communication between one or more Data Source Connectors and IBM backup service, as well as the IBM Cloud Storage.
| Source | Destination | Port | Protocol | Purpose |
|---|---|---|---|---|
| Data Source Connector | IBM Backup Service | 443 | TCP | Required for Backup and Recovery operations. |
| Data Source Connector | IBM Backup Service | 29991 | TCP | Required for Backup and Recovery operations. |
| Data Source Connector | IBM Cloud Storage | 443 | TCP | Required for Backup and Recovery operations. |
Others
| Source | Destination | Port | Protocol | Purpose |
|---|---|---|---|---|
| Data Source Connector | Internet/Internal | 123,323 | UDP | Required for time.google.com OR internal NTP |
| Data Source Connector | Internet/Internal | 53 | TCP, UDP | Required for DNS or internal DNS |
Create a Data Source Connection in VMware
To create a Data Source Connection:
-
In IBM Cloud Backup and Recovery, navigate to
System>Data Source Connections. -
Click
New Connection. -
In the
Create Data Source Connectiondialog, select the following:-
From the
Deployment Platformdrop-down, selectVMware. -
To deploy the Data Source Connector in your data center, click
Copy OVA URL. The OVA URL is used to deploy the OVA template in VMware vCenter or ESXi. -
Copy the
Connection tokenand clickCreate. The Connection token is used to link or claim the Data Source Connector with the created connection.
-
-
To deploy the Data Source Connector OVA in your data center:
-
Log in to your vCenter host.
-
From the
Hosts and Clusterstab in the vSphere Web Client, right-click on any cluster that can host your VM and selectDeploy OVF Template. The Deploy OVF Template wizard opens. -
On the
Select an OVF templatepage, do one of the following and click Next:- Paste the link of the OVA file that you copied in Step 3 (b) in the
URLfield. - Select
Local file, clickUPLOAD FILES, and browse to the location of the OVA file you downloaded in step 3 (b).
- Paste the link of the OVA file that you copied in Step 3 (b) in the
-
On the
Select a name and folderpage, enter the following and clickNext:- In the
Virtual machine namefield, enter a unique name for your Data Source Connector. - In the
Select a location for the Virtual Machinefield, select where your VM should reside from the displayed list of inventory locations.
- In the
-
On the
Compute Resourcespage, select a compute resource for the Data Source Connector VM and clickNext. -
On the
Review detailspage, verify the Data Source Connector information and clickNext. -
On the
Configurationpage, verifySAAS-CONNECTORis selected and clickNext. -
On the
Select storagepage, select a datastore with at least 171 GB of free disk space and clickNext. -
On the
Select networkspage, select a destination network and clickNext. You can select VLANs from both theDataNetworkand theSecondaryNetworkfields. The Data Network is used for communication with Data Source, and the Secondary Network is used for communication with your data sources. Based on your requirements:- To deploy the Data Source Connector on a single network, select the same VLAN in both
-
DataNetwork and SecondaryNetwork
-
To deploy the Data Source Connector on a dual network, select different VLANs in
DataNetworkandSecondaryNetwork, respectively. -
The Data Source Connector must have dual IP addresses if your data sources are in a private non-routable VLAN.
-
Once you have deployed the Data Source Connector on a single network, you cannot modify the Data Source Connector to use a dual network or vice versa.
- On the
Customize templatepage, enter the network settings:Network IP Address,Network Netmask, andDefault Gateway. If you have selected a different VLAN for the secondary network, enter theNetwork IP Address,Network Netmask, andDefault Gatewayfor the secondary network, as well. ClickNext.
- On the
-
To set the network settings using static IP addresses, manually enter the details in the respective fields for both DataNetwork and SecondaryNetwork.
-
To set the network settings using DHCP, leave the fields blank in both the DataNetwork and SecondaryNetwork sections.
-
Data Network and Secondary Network must be configured using the same network configuration method. That is static IP addresses or DHCP.
-
Review the summary on the
Ready to completepage and clickFinish. -
Once the VM is created, power it on. After it boots, the services in the Data Source Connector VM (including the UI) can take 4-5 minutes to start.
-
-
Enter the IP address of the Data Source Connector VM in the address bar of your browser and click
Enter. -
On the Data Source Connector's User Interface, enter
adminin theUsernameandPasswordfields to log in to the Data Source Connector.
On the next screen, you are prompted to change your password. Change your default password and log in again with your new password.
-
Verify the network configuration settings, make necessary changes, and click
Continue. -
On the
DataSourceConnectorConfigurationpage, paste theConnectiontokenin theConnectionClaimTokenfield and clickSave.
It can take another few minutes for the Data Source Connector to authenticate to the IBM Cloud Backup and Recovery Service. Click Data Source Connection to list the Data Source Connector(s) that are claimed.
Create a Data Source Connection in VPC
To create a Data Source Connection:
-
In IBM Cloud Backup and Recovery, navigate to
System>Data Source Connections. -
Click
New Connection. -
In the
Create Data Source Connectiondialog, select the following:a. From the
Deployment Platformdrop-down, selectVPC.- NOTE: The VPC option will not be available at launch. It is ok to leave the Deployment Platform as VMware
b. Copy the
Connection tokenand clickCreate. The Connection token is utilized to link or claim the Data Source Connector with the created connection.c. To deploy the Data Source Connector in your VPC, Go to the IBM Cloud Catalog
d. Search for
Backup and Recoverye. Select the
Backup and Recovery Data Source Connectorimage and click the catalog tile- NOTE: If the image mentioned earlier is not available, the User might have to accept an invite. An invite will have been sent to the account admin's email or can be seen in
https://cloud.ibm.com/notificationsby the admin.
f. Accept the terms and select
Continueh. TheVirtual server for VPCcreate page opens, Select appropriateLocation,Name,Resource Group,VPCand create the Virtual Service.- NOTE: The
Data Source Connectorneeds to have access to the intended workloads. The recommendation is to create it within the sameVPCas the workloads.
Once the VSI is created and powered on, the services in the Data Source Connector VM (including the UI) can take 4-5 minutes to start.
The initial VSI deployment is your responsibility, after that the OS patching and version upgrades are managed by the Backup and Recovery service.
-
Enter the IP address of the Data Source Connector VM in the address bar of your browser and click
Enter. -
On the Data Source Connector's User Interface, enter
adminin theUsernameandPasswordfields to log in to the Data Source Connector.
On the next screen, you are prompted to change your password. Change your default password and log in again with your new password.
-
Verify the network configuration settings, make necessary changes, and click
Continue. -
On the
DataSourceConnectorConfigurationpage, paste theConnectiontokenin theConnectionClaimTokenfield and clickSave.
It can take another few minutes for the Data Source Connector to authenticate to the IBM Cloud Backup and Recovery Service. Click Data Source Connection to list the Data Source Connector(s) that are claimed.
VPE Gateways
While the VSI can connect to the Backup and Recovery instance, a VPE gateway provides a better performance, To create a VPE gateway, follow these steps.
- From the IBM Cloud catalog, search for
Virtual private endpointand clickVirtual private endpoint for VPCtile. - Select the appropriate
Location,Name,Resource Group,VPC. - Under
Request connection to a service, SelectIBM Cloud Service. - Select
Cloud service offeringsto beBackup and Recoveryand the appropriateCloud service regions. - Select the appropriate
Backup and Serviceinstance in the table below. - Configure the
Reserved IPas required and create the VPE Gateway.
Create a data source connection for Kubernetes/OpenShift
To create a data source connection for Kubernetes or OpenShift clusters, including detailed guidance on choosing the correct deployment platform and managing connections, see Create a data source connection in the Kubernetes/OpenShift getting started guide.
The process involves:
- Selecting the appropriate deployment platform (ROKS VPC, IKS VPC, ROKS classic, or IKS classic) that matches your cluster infrastructure
- Creating the connection in the IBM Cloud Backup and Recovery dashboard
- Copying the provided Helm install command for use in the next step
You can reuse an existing connection for multiple clusters on the same deployment platform.
Install Data Source Connector on Kubernetes and OpenShift
The Data Source Connector is deployed as a StatefulSet with 2 replicas (by default) on your Kubernetes or OpenShift cluster. This establishes the communication channel between your cluster and the IBM Cloud Backup and Recovery service.
For clusters with private endpoints only: You must run ibmcloud, kubectl and helm commands from IBM Cloud Shell.
For clusters with public endpoints: You can run ibmcloud, kubectl, and helm commands from either IBM Cloud Shell or your local workspace.
Resource requirements
Ensure that your cluster has sufficient CPU and memory resources. The Data Source Connector will consume resources from your cluster nodes. Additional backup agent components (Datamover and Velero) will be deployed later during source registration.
Data Source Connector resource consumption:
| Component | Deployment Type | CPU Requests | Memory Requests | Total (default 2 replicas) |
|---|---|---|---|---|
| Data Source Connector | StatefulSet (2 replicas) | 2 per replica | 5Gi per replica | 4 CPU, 10Gi Memory |
Note: Datamover (DaemonSet) and Velero components are deployed automatically during source registration and are not part of the initial Data Source Connector installation.
-
Open IBM Cloud Shell (or use your local workspace if your cluster has a public endpoint).
-
Identify the source cluster where you want to install the Data Source Connector. This should be the Kubernetes or OpenShift cluster that you want to back up and protect. You must have admin access to this cluster to install the Data Source Connector.
List the available clusters:
ibmcloud ks cluster lsFrom the output, note the cluster name where the Data Source Connector should be deployed. Ensure you have admin privileges for this cluster.
-
Download and configure the
KUBECONFIGfor the selected cluster with admin privileges:ibmcloud ks cluster config --cluster <cluster-name> --admin -
The Helm chart is hosted in the IBM Container Registry (ICR). Log in to the Helm/OCI registry by using the following command:
helm registry login icr.io --username iamapikey --password "${API_KEY}"See Creating an API_KEY to create a new API_KEY if you don't have an existing one.
-
Retrieve the Helm install command that you copied earlier in the Create a data source connection section and update it based on your cluster type.
For IKS Classic clusters:
You must specify a storage class that is available on Classic clusters and disable SCC (Security Context Constraints) as it's specific to OpenShift:
helm upgrade --install <k8-app-name> oci://icr.io/ext/brs/brs-ds-connector-chart --version 7.2.18-release-20260226-49768040 --set secrets.registrationToken=<your-registration-token> --set deploymentPlatform.rocp.sccEnabled=false --set volumeClaimTemplate.storageClass=ibmc-block-bronze --namespace ibm-brs-data-source-connector --create-namespaceFor IKS VPC clusters:
The default storage class
ibmc-vpc-block-metro-5iops-tieris used automatically. Disable SCC as it's specific to OpenShift:helm upgrade --install <k8-app-name> oci://icr.io/ext/brs/brs-ds-connector-chart --version 7.2.18-release-20260226-49768040 --set secrets.registrationToken=<your-registration-token> --set deploymentPlatform.rocp.sccEnabled=false --namespace ibm-brs-data-source-connector --create-namespaceFor ROKS Classic clusters:
You must specify a storage class that is available on Classic clusters. SCC is enabled by default for OpenShift:
helm upgrade --install <k8-app-name> oci://icr.io/ext/brs/brs-ds-connector-chart --version 7.2.18-release-20260226-49768040 --set secrets.registrationToken=<your-registration-token> --set volumeClaimTemplate.storageClass=ibmc-block-bronze --namespace ibm-brs-data-source-connector --create-namespaceFor ROKS VPC clusters:
The default storage class
ibmc-vpc-block-metro-5iops-tieris used automatically. SCC is enabled by default for OpenShift:helm upgrade --install <k8-app-name> oci://icr.io/ext/brs/brs-ds-connector-chart --version 7.2.18-release-20260226-49768040 --set secrets.registrationToken=<your-registration-token> --namespace ibm-brs-data-source-connector --create-namespaceThe default storage class
ibmc-vpc-block-metro-5iops-tieris only available on VPC clusters and will not work on Classic clusters. For Classic clusters, useibmc-block-bronze,ibmc-block-silver, oribmc-block-gold. -
Run the appropriate Helm install command in the IBM Cloud Shell based on your cluster type.
-
Check that the Helm release is installed:
helm list -n ibm-brs-data-source-connectorCheck the data source connector pods status:
kubectl get pods -n ibm-brs-data-source-connectorCustomizing the Helm install command
You can customize the Helm install command with the following optional flags:
-
--namespace <namespace-name>: Specifies the namespace where the Data Source Connector will be deployed. If not specified, the default namespace is used. -
--create-namespace: Creates the specified namespace if it doesn't exist. This flag is useful when deploying to a new namespace. -
--set volumeClaimTemplate.storageClass=<storage-class-name>: Specifies the StorageClass to use for provisioning the persistent volume for the Data Source Connector. This parameter sets the storage class specifically for the Data Source Connector's persistent volume. This is required for Classic clusters as the default VPC storage class (ibmc-vpc-block-metro-5iops-tier) cannot be used on Classic clusters. If you want to use VPC File CSI driver storage classes, you need to create a custom storage class withGID:1000andUID:1000because the Data Source Connector runs as non-root app. For reference on creating storage classes, see Deploying an app that runs as non-root. -
--set replicaCount=<number>: Sets the number of Data Source Connector pod replicas. The default value is 3. Adjust this based on your high availability and workload requirements. -
--set fullnameOverride=<name>: Assigns a specific name to the Data Source Connector pods, making them easier to identify and manage. -
--set deploymentPlatform.rocp.sccEnabled=false: Disables Security Context Constraints (SCC) for the deployment. This flag should be used for IBM Kubernetes Service clusters, as SCC is specific to OpenShift clusters. -
--set nodeSelector.<key>="<value>": Schedules Data Source Connector pods on nodes with the specified label. This is useful when you want to run the connector on dedicated worker nodes. Use the default worker pool labelibm-cloud.kubernetes.io/worker-pool-name=<worker-pool-name>to target a specific worker pool. -
--set "tolerations[<index>].key=<key>": Sets the toleration key for node taints. Tolerations allow pods to be scheduled on nodes with matching taints. -
--set "tolerations[<index>].operator=<operator>": Specifies the toleration operator. Common values areEqual(exact match) orExists(key exists). -
--set "tolerations[<index>].value=<value>": Sets the toleration value that must match the taint value on the node. -
--set "tolerations[<index>].effect=<effect>": Specifies the taint effect. Common values areNoSchedule(prevents scheduling),PreferNoSchedule(tries to avoid scheduling), orNoExecute(evicts existing pods).
Adding a dedicated worker pool for Data Source Connector
It's recommended to create a dedicated worker pool for the Data Source Connector. This helps ensure that the connector pods run on a dedicated worker pool with appropriate taints to ensure workload isolation and performance.
-
Deploy a dedicated worker pool. All worker pools automatically have the label
ibm-cloud.kubernetes.io/worker-pool-name=<worker-pool-name>which can be used for node selection. -
Add taints to the worker pool:
ibmcloud ks worker-pool taint set \ --cluster "${CLUSTER_NAME_ID}" \ --worker-pool "${WORKER_POOL_NAME}" \ --taint "dedicated=data-source-connector:NoSchedule" -f -
Deploy the Data Source Connector to the dedicated worker pool.
Example Helm install command:
helm upgrade --install dsc-test oci://icr.io/ext/brs/brs-ds-connector-chart \ --version 7.2.18-release-20260226-49768040 \ --namespace ibm-brs-data-source-connector \ --create-namespace \ --set secrets.registrationToken=xxx \ --set fullnameOverride=dsc \ --set replicaCount=2 \ --set nodeSelector.ibm-cloud\\.kubernetes\\.io/worker-pool-name="my-worker-pool" \ --set "tolerations[0].key=dedicated" \ --set "tolerations[0].operator=Equal" \ --set "tolerations[0].value=data-source-connector" \ --set "tolerations[0].effect=NoSchedule"Replace
my-worker-poolwith your actual worker pool name. -
Upgrading the Data Source Connector
When new releases of the IBM Cloud Backup and Recovery service become available, there is a possibility that a new Data Source Connector version is also available. It is recommended to upgrade your Data Source Connector when new releases become available to ensure you have the latest features, security patches, and bug fixes.
To check for new service releases, see the IBM Cloud Backup and Recovery release notes.
Upgrades for the Data Source Connector are currently manual. Follow these steps to check your current version and upgrade to a newer version:
Check your current Data Source Connector version
To find the currently installed version of the Data Source Connector:
helm list -n ibm-brs-data-source-connector
This command displays the installed Helm releases in the ibm-brs-data-source-connector namespace, including the chart version (APP VERSION column).
Alternatively, you can check the version using kubectl:
kubectl get statefulset -n ibm-brs-data-source-connector -o jsonpath='{.items[0].spec.template.spec.containers[0].image}'
This command shows the container image tag, which includes the version number.
Check for available versions
To view all available versions of the Data Source Connector chart, first log in to your IBM Cloud account:
ibmcloud login -a cloud.ibm.com --apikey "${API_KEY}"
Then, log in to the IBM Container Registry:
ibmcloud cr login
After logging in, list the available Data Source Connector images:
ibmcloud cr images --restrict ext/brs/brs-ds-connector
This command displays all available Data Source Connector versions. Compare the output with your currently installed version to determine if an upgrade is available.
Upgrade to a newer version
Once you've identified the target version, use the Helm upgrade command. You must provide the registration token because the old token may have expired.
helm upgrade --install <release-name> oci://icr.io/ext/brs/brs-ds-connector-chart --version <new-version> --reuse-values --set secrets.registrationToken=<token> -n ibm-brs-data-source-connector
Replace <release-name> with your actual Helm release name (use helm list -n ibm-brs-data-source-connector to find it), <new-version> with the target version number, and <token> with the registration token obtained from the Data Source Connection.
To obtain the registration token:
- Log in to your IBM Cloud Backup and Recovery instance in the IBM Cloud console.
- Navigate to
System>Data Source Connections. - Locate the Data Source Connection that corresponds to your cluster.
- Click on the three dots (⋮) menu corresponding to the connection.
- Select
Connection Tokenfrom the menu options. - Copy the displayed connection token (also referred to as the registration token).
- Use this token in the
secrets.registrationTokenparameter of the upgrade command.
You must provide the registration token in the upgrade command because the previous token may have expired.
If you need to create a new Data Source Connection, see Create a data source connection for Kubernetes/OpenShift.
The --reuse-values flag preserves your existing configuration settings during the upgrade.
Example upgrade command:
helm upgrade --install dsc-test oci://icr.io/ext/brs/brs-ds-connector-chart --version 7.2.19-release-20260301-12345678 --reuse-values --set secrets.registrationToken=<new-token> -n ibm-brs-data-source-connector
Verify the upgrade
After upgrading, verify that the new version is running:
helm list -n ibm-brs-data-source-connector
Check the pod status to ensure all pods are running with the new version:
kubectl get pods -n ibm-brs-data-source-connector
Alternative: Using a values.yaml file
As an alternative to using multiple --set flags, you can use a values.yaml file to pass configuration values. This approach is cleaner and easier to manage, especially when dealing with complex configurations.
-
Create a
custom-values.yamlfile with the following content:secrets: registrationToken: "<your-registration-token>" nodeSelector: ibm-cloud.kubernetes.io/worker-pool-name: "<your-worker-pool-name>" tolerations: - key: "dedicated" operator: "Equal" value: "data-source-connector" effect: "NoSchedule" -
Run the Helm install command using the values file:
helm upgrade --install dsc oci://icr.io/ext/brs/brs-ds-connector-chart \ --version 7.2.18-release-20260226-49768040 \ --namespace ibm-brs-data-source-connector \ --create-namespace \ -f custom-values.yaml