8.2 to 8.2.0.1
1. Upgrade From 8.2 to 8.2.0.1(Selected Services)
RDAF Platform: From 8.2 to 8.2.0.1(Selected Services)
AIOps (OIA) Application: From 8.2 to 8.2.0.1(Selected Services)
RDAF Deployment rdafk8s CLI: From 1.5.0 to 1.5.0.1
RDAF Client rdac CLI: From 8.2 to 8.2.0.1
RDAF Studio: From 8.2 to 8.2.0.1
RDAF Platform: From 8.2 to 8.2.0.1(Selected Services)
OIA (AIOps) Application: From 8.2 to 8.2.0.1(Selected Services)
RDAF Deployment rdaf CLI: From 1.5.0 to 1.5.0.1
RDAF Client rdac CLI: From 8.2 to 8.2.0.1
RDAF Studio: From 8.2 to 8.2.0.1
1.1. Prerequisites
Before proceeding with this upgrade, please make sure and verify the below prerequisites are met.
Currently deployed CLI and RDAF services are running the below versions.
-
RDAF Deployment CLI version: 1.5.0
-
Infra Services tag: 1.0.4
-
Platform Services and RDA Worker tag: 8.2
-
OIA Application Services tag: 8.2
-
CloudFabrix recommends taking VMware VM snapshots where RDA Fabric infra/platform/applications are deployed
Note
- Check the Disk space of all the Platform and Service Vm's using the below mentioned command, the highlighted disk size should be less than 80%
rdauser@oia-125-216:~/collab-3.7-upgrade$ df -kh
Filesystem Size Used Avail Use% Mounted on
udev 32G 0 32G 0% /dev
tmpfs 6.3G 357M 6.0G 6% /run
/dev/mapper/ubuntu--vg-ubuntu--lv 48G 12G 34G 26% /
tmpfs 32G 0 32G 0% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 32G 0 32G 0% /sys/fs/cgroup
/dev/loop0 64M 64M 0 100% /snap/core20/2318
/dev/loop2 92M 92M 0 100% /snap/lxd/24061
/dev/sda2 1.5G 309M 1.1G 23% /boot
/dev/sdf 50G 3.8G 47G 8% /var/mysql
/dev/loop3 39M 39M 0 100% /snap/snapd/21759
/dev/sdg 50G 541M 50G 2% /192.168-data
/dev/loop4 92M 92M 0 100% /snap/lxd/29619
/dev/loop5 39M 39M 0 100% /snap/snapd/21465
/dev/sde 15G 140M 15G 1% /zookeeper
/dev/sdd 30G 884M 30G 3% /kafka-logs
/dev/sdc 50G 3.3G 47G 7% /opt
/dev/sdb 50G 29G 22G 57% /var/lib/docker
/dev/sdi 25G 294M 25G 2% /graphdb
/dev/sdh 50G 34G 17G 68% /opensearch
/dev/loop6 64M 64M 0 100% /snap/core20/2379
- Check all MariaDB nodes are sync on HA setup using below commands before start upgrade
Tip
Please run the below commands on the VM host where RDAF deployment CLI was installed and rdafk8s setup command was run. The mariadb configuration is read from /opt/rdaf/rdaf.cfg file.
MARIADB_HOST=`cat /opt/rdaf/rdaf.cfg | grep -A3 haproxy| grep advertised_external_host | awk '{print $3}'`
MARIADB_USER=`cat /opt/rdaf/rdaf.cfg | grep -A3 mariadb | grep user | awk '{print $3}' | base64 -d`
MARIADB_PASSWORD=`cat /opt/rdaf/rdaf.cfg | grep -A3 mariadb | grep password | awk '{print $3}' | base64 -d`
mysql -u$MARIADB_USER -p$MARIADB_PASSWORD -h $MARIADB_HOST -P3307 -e "show status like 'wsrep_local_state_comment';"
Please verify that the mariadb cluster state is in Synced state.
+---------------------------+--------+
| Variable_name | Value |
+---------------------------+--------+
| wsrep_local_state_comment | Synced |
+---------------------------+--------+
Please run the below command and verify that the mariadb cluster size is 3.
mysql -u$MARIADB_USER -p$MARIADB_PASSWORD -h $MARIADB_HOST -P3307 -e "SHOW GLOBAL STATUS LIKE 'wsrep_cluster_size'";
+--------------------+-------+
| Variable_name | Value |
+--------------------+-------+
| wsrep_cluster_size | 3 |
+--------------------+-------+
Warning
Make sure all of the above pre-requisites are met before proceeding with the upgrade process.
Warning
Kubernetes: Though Kubernetes based RDA Fabric deployment supports zero downtime upgrade, it is recommended to schedule a maintenance window for upgrading RDAF Platform and AIOps services to newer version.
Important
Please make sure full backup of the RDAF platform system is completed before performing the upgrade.
Kubernetes: Please run the below backup command to take the backup of application data.
Run the below command on RDAF Management system and make sure the Kubernetes PODs are NOT in restarting mode (it is applicable to only Kubernetes environment)
- Verify that RDAF deployment
rdafcli version is 1.5.0 on the VM where CLI was installed for docker on-prem registry managing Kubernetes or Non-kubernetes deployments.
- On-premise docker registry service version is 1.0.4
0889e08f0871 docker2.cloudfabrix.io:443/external/docker-registry:1.0.4 "/entrypoint.sh /bin…" 7 days ago Up 7 days deployment-scripts-docker-registry-1
-
RDAF Infrastructure services version is 1.0.4 except for below services.
-
rda-minio: version is
RELEASE.2024-12-18T13-15-44Z
Run the below command to get rdafk8s Infra service details
+--------------------------+-----------------+-------------------+--------------+--------------------------------+
| Name | Host | Status | Container Id | Tag |
+--------------------------+-----------------+-------------------+--------------+--------------------------------+
| rda-nats | 192.168.108.114 | Up 19 Minutes ago | bbb50d2dacc5 | 1.0.4 |
| rda-minio | 192.168.108.114 | Up 19 Minutes ago | d26148d4bf44 | RELEASE.2024-12-18T13-15-44Z |
| rda-mariadb | 192.168.108.114 | Up 19 Minutes ago | 02975e0eec89 | 1.0.4 |
| rda-opensearch | 192.168.108.114 | Up 18 Minutes ago | 1494be76f694 | 1.0.4 |
+--------------------------+-----------------+-------------------+--------------+--------------------------------+
- RDAF Platform services version is 8.2
Run the below command to get RDAF Platform services details
+---------------+-----------------+---------------+--------------+---------+
| Name | Host | Status | Container Id | Tag |
+---------------+-----------------+---------------+--------------+---------+
| rda-api- | 192.168.108.119 | Up 14 Hours | 2ca4370a175a | 8.2 |
| server | | ago | | |
| rda-api- | 192.168.108.120 | Up 14 Hours | cce0d6bcba36 | 8.2 |
| server | | ago | | |
| rda-registry | 192.168.108.120 | Up 14 Hours | e029a9ff96fe | 8.2 |
| | | ago | | |
| rda-registry | 192.168.108.119 | Up 14 Hours | eacbc82ae8c9 | 8.2 |
| | | ago | | |
| rda-identity | 192.168.108.120 | Up 14 Hours | 45409c977c7c | 8.2 |
| | | ago | | |
| rda-identity | 192.168.108.119 | Up 14 Hours | 584458932e2c | 8.2 |
+---------------+-----------------+---------------+--------------+---------+
- RDAF Worker version is 8.2
Run the below command to get RDAF Worker details
+------------+----------------+------------+--------------+---------+
| Name | Host | Status | Container Id | Tag |
+------------+----------------+------------+--------------+---------+
| rda_worker | 192.168.125.63 | Up 7 weeks | cfe1fe65c692 | 8.2 |
+------------+----------------+------------+--------------+---------+
- RDAF OIA Application services version is 8.2
Run the below command to get RDAF App services details
+-------------------------------+-----------------+-----------------+--------------+-------+
| Name | Host | Status | Container Id | Tag |
+-------------------------------+-----------------+-----------------+--------------+-------+
| rda-alert-correlator | 192.168.108.118 | Up 14 Hours ago | afdbbe6453e4 | 8.2 |
| rda-alert-correlator | 192.168.108.117 | Up 14 Hours ago | 631b7978dcb0 | 8.2 |
| rda-alert-ingester | 192.168.108.117 | Up 14 Hours ago | 33322e0b9cb9 | 8.2 |
| rda-alert-ingester | 192.168.108.118 | Up 14 Hours ago | 8178c043bd04 | 8.2 |
| rda-alert-processor | 192.168.108.117 | Up 14 Hours ago | b342b582ea1d | 8.2 |
| rda-alert-processor | 192.168.108.118 | Up 14 Hours ago | b6f85413c2df | 8.2 |
+-------------------------------+-----------------+-----------------+--------------+-------+
Currently deployed CLI and RDAF services are running the below versions.
-
RDAF Deployment CLI version: 1.5.0
-
Infra Services tag: 1.0.4
-
Platform Services and RDA Worker tag: 8.2
-
OIA Application Services tag: 8.2
-
CloudFabrix recommends taking VMware VM snapshots where RDA Fabric infra/platform/applications are deployed
Note
- Check the Disk space of all the Platform and Service Vm's using the below mentioned command, the highlighted disk size should be less than 80%
rdauser@oia-125-216:~/collab-3.7-upgrade$ df -kh
Filesystem Size Used Avail Use% Mounted on
udev 32G 0 32G 0% /dev
tmpfs 6.3G 357M 6.0G 6% /run
/dev/mapper/ubuntu--vg-ubuntu--lv 48G 12G 34G 26% /
tmpfs 32G 0 32G 0% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 32G 0 32G 0% /sys/fs/cgroup
/dev/loop0 64M 64M 0 100% /snap/core20/2318
/dev/loop2 92M 92M 0 100% /snap/lxd/24061
/dev/sda2 1.5G 309M 1.1G 23% /boot
/dev/sdf 50G 3.8G 47G 8% /var/mysql
/dev/loop3 39M 39M 0 100% /snap/snapd/21759
/dev/sdg 50G 541M 50G 2% /192.168-data
/dev/loop4 92M 92M 0 100% /snap/lxd/29619
/dev/loop5 39M 39M 0 100% /snap/snapd/21465
/dev/sde 15G 140M 15G 1% /zookeeper
/dev/sdd 30G 884M 30G 3% /kafka-logs
/dev/sdc 50G 3.3G 47G 7% /opt
/dev/sdb 50G 29G 22G 57% /var/lib/docker
/dev/sdi 25G 294M 25G 2% /graphdb
/dev/sdh 50G 34G 17G 68% /opensearch
/dev/loop6 64M 64M 0 100% /snap/core20/2379
Warning
Make sure all of the above pre-requisites are met before proceeding with the upgrade process.
Warning
Non-Kubernetes: Upgrading RDAF Platform and AIOps application services is a disruptive operation. Schedule a maintenance window before upgrading RDAF Platform and AIOps services to newer version.
Important
Please make sure full backup of the RDAF platform system is completed before performing the upgrade.
Non-Kubernetes: Please run the below backup command to take the backup of application data.
Note: Please make sure this backup-dir is mounted across all infra,cli vms.- Verify that RDAF deployment
rdafcli version is 1.5.0 on the VM where CLI was installed for docker on-prem registry managing Kubernetes or Non-kubernetes deployments.
- On-premise docker registry service version is 1.0.4
173d38eebeab docker2.cloudfabrix.io:443/external/docker-registry:1.0.4 "/entrypoint.sh /bin…" 45 hours ago Up 45 hours deployment-scripts-docker-registry-1
-
RDAF Infrastructure services version is 1.0.4 except for below services.
-
rda-minio: version is
RELEASE.2024-12-18T13-15-44Z
Run the below command to get RDAF Infra service details
+-------------------+----------------+-------------+--------------+------------------------------+
| Name | Host | Status | Container Id | Tag |
+-------------------+----------------+-------------+--------------+------------------------------+
| nats | 192.168.125.63 | Up 2 months | aff2eb1f37c9 | 1.0.4 |
| minio | 192.168.125.63 | Up 2 months | ed6bb3ea036a | RELEASE.2024-12-18T13-15-44Z |
| mariadb | 192.168.125.63 | Up 2 months | 616a98d6471c | 1.0.4 |
| opensearch | 192.168.125.63 | Up 2 months | 7edeede52a9b | 1.0.4 |
| kafka | 192.168.125.63 | Up 2 months | d1426429da4c | 1.0.4 |
| graphdb[operator] | 192.168.125.63 | Up 2 months | 8a53795f6ee4 | 1.0.4 |
| graphdb[server] | 192.168.125.63 | Up 2 months | 06c187c7dfa2 | 1.0.4 |
| haproxy | 192.168.125.63 | Up 2 months | fde40536be0c | 1.0.4 |
+-------------------+----------------+-------------+--------------+------------------------------+
- RDAF Platform services version is 8.2
Run the below command to get RDAF Platform services details
+--------------------------+----------------+------------+--------------+-------+
| Name | Host | Status | Container Id | Tag |
+--------------------------+----------------+------------+--------------+-------+
| rda_api_server | 192.168.125.63 | Up 7 weeks | c6500e23738f | 8.2 |
| rda_registry | 192.168.125.63 | Up 7 weeks | 34f008691fd4 | 8.2 |
| rda_scheduler | 192.168.125.63 | Up 7 weeks | 8b358f65a7d3 | 8.2 |
| rda_collector | 192.168.125.63 | Up 7 weeks | 1888441693c0 | 8.2 |
| rda_identity | 192.168.125.63 | Up 7 weeks | 10e43ae93430 | 8.2 |
| rda_asm | 192.168.125.63 | Up 7 weeks | f98c2c79539a | 8.2 |
+--------------------------+----------------+------------+--------------+-------+
- RDAF Worker version is 8.2
Run the below command to get RDAF Worker details
+------------+----------------+------------+--------------+-------+
| Name | Host | Status | Container Id | Tag |
+------------+----------------+------------+--------------+-------+
| rda_worker | 192.168.125.63 | Up 7 weeks | bc46556f64d2 | 8.2 |
+------------+----------------+------------+--------------+-------+
- RDAF OIA Application services version is 8.2
Run the below command to get RDAF App services details
+-----------------------------------+----------------+------------+--------------+-------+
| Name | Host | Status | Container Id | Tag |
+-----------------------------------+----------------+------------+--------------+-------+
| cfx-rda-app-controller | 192.168.125.63 | Up 7 weeks | 1bae5abb4e9c | 8.2 |
| cfx-rda-reports-registry | 192.168.125.63 | Up 7 weeks | 925a97ecb0a3 | 8.2 |
| cfx-rda-notification-service | 192.168.125.63 | Up 7 weeks | 1628da0a7a30 | 8.2 |
| cfx-rda-file-browser | 192.168.125.63 | Up 7 weeks | 237c85c6cb9f | 8.2 |
| cfx-rda-configuration-service | 192.168.125.63 | Up 7 weeks | 0fe8f3ee7596 | 8.2 |
| cfx-rda-alert-ingester | 192.168.125.63 | Up 7 weeks | d58452342e72 | 8.2 |
| cfx-rda-webhook-server | 192.168.125.63 | Up 7 weeks | f3578f725d9c | 8.2 |
+-----------------------------------+----------------+------------+--------------+-------+
RDAF Deployment CLI Upgrade:
Please follow the below given steps.
Note
Upgrade RDAF Deployment CLI on both on-premise docker registry VM and RDAF Platform's management VM if provisioned separately.
Login into the VM where rdaf deployment CLI was installed for docker on-premise registry and managing Kubernetes or Non-kubernetes deployment.
- Download the RDAF Deployment CLI's newer version 1.5.0.1 bundle.
wget https://macaw-amer.s3.us-east-1.amazonaws.com/releases/rdaf-platform/1.5.0.1/rdafcli-1.5.0.1.tar.gz
- Upgrade the
rdafk8sCLI to version 1.5.0.1
- Verify the installed
rdafk8sCLI version is upgraded to 1.5.0.1
- Download the RDAF Deployment CLI's newer version 1.5.0.1 bundle and copy it to RDAF CLI management VM on which
rdafdeployment CLI was installed.
- Download the RDAF Deployment CLI's newer version 1.5.0.1 bundle
wget https://macaw-amer.s3.us-east-1.amazonaws.com/releases/rdaf-platform/1.5.0.1/rdafcli-1.5.0.1.tar.gz
- Upgrade the
rdafCLI to version 1.5.0.1
- Verify the installed
rdafCLI version is upgraded to 1.5.0.1
- Download the RDAF Deployment CLI's newer version 1.5.0.1 bundle and copy it to RDAF management VM on which
rdaf & rdafk8sdeployment CLI was installed.
1.2. Upgrade Steps
Please download the below given python script
Note
This is applicable only in environments where Agentic AI is deployed and specifically for HA deployments.
Please run the below python upgrade script
Please download the below python script (rdaf_upgrade_150_1501.py)
wget https://macaw-amer.s3.us-east-1.amazonaws.com/releases/rdaf-platform/1.5.0.1/rdaf_upgrade_150_1501.py
Please run the downloaded python upgrade script rdaf_upgrade_150_1501.py as shown below
Note
The above command will show the available options for the upgrade script
rdauser@infra108122:~$ python rdaf_upgrade_150_1501.py -h
usage: rdaf_upgrade_150_1501.py [-h] {upgrade} ...
options:
-h, --help show this help message and exit
options:
{upgrade} Available options
upgrade upgrade the setup
This below given command will execute the upgrade process after user has downloaded the script and reviewed the available options with the help command.
Portal token updated for user creation (sub=portal_token).
Admin user already exists; platform UI is accessible.
Processing nginx.conf on host 192.168.108.114
/opt/rdaf/config/nginx_config/nginx.conf does not exist, skipping changes
Processing nginx.conf on host 192.168.108.115
/opt/rdaf/config/nginx_config/nginx.conf does not exist, skipping changes
Please download the below given python script
Note
This is applicable only in environments where Agentic AI is deployed and specifically for HA deployments.
Please run the below python upgrade script
Creating backup of existing haproxy.cfg on host 192.168.133.60
Restarting HAProxy on host: 192.168.133.60
Container infra-haproxy-1 Restarting
Container infra-haproxy-1 Started
HAProxy restarted successfully on 192.168.133.60
Creating backup of existing haproxy.cfg on host 192.168.133.61
Restarting HAProxy on host: 192.168.133.61
[+] Restarting 1/1
✔ Container infra-haproxy-1 Started 10.6s
HAProxy restarted successfully on 192.168.133.61
Please download the below python script (rdaf_upgrade_150_1501.py)
wget https://macaw-amer.s3.us-east-1.amazonaws.com/releases/rdaf-platform/1.5.0.1/rdaf_upgrade_150_1501.py
Please run the downloaded python upgrade script rdaf_upgrade_150_1501.py as shown below
Note
The above command will show the available options for the upgrade script
rdauser@infra108122:~$ python rdaf_upgrade_150_1501.py -h
usage: rdaf_upgrade_150_1501.py [-h] {upgrade} ...
options:
-h, --help show this help message and exit
options:
{upgrade} Available options
upgrade upgrade the setup
This below given command will execute the upgrade process after user has downloaded the script and reviewed the available options with the help command.
Note
When an infra node goes down and the VM is restarted, GraphDB does not start automatically. The script below addresses this issue by ensuring GraphDB starts properly.
Portal token updated for user creation (sub=portal_token).
Admin user already exists; platform UI is accessible.
Processing nginx.conf on host 192.168.133.60
/opt/rdaf/config/nginx_config/nginx.conf does not exist, skipping changes
Processing nginx.conf on host 192.168.133.61
/opt/rdaf/config/nginx_config/nginx.conf does not exist, skipping changes
1.2.1 Download the new Docker Images
Login into the VM where rdaf deployment CLI was installed for docker on-premise registry and managing kubernetes & Non-kubernetes deployment.
Download the new docker image tags for RDAF Platform and OIA (AIOps) Application services and wait until all of the images are downloaded.
Note
If the download of the images fail, Please re-execute the above command
Run the below command to verify above mentioned tags are downloaded for all of the RDAF Platform and OIA (AIOps) Application services.
Please make sure 8.2.0.1 image tag is downloaded for the below RDAF Platform services.
- rda-client-api-server
- rda-scheduler
- rda-collector
- onprem-portal
- rda-chat-helper
- rda-resource-manager
- rda-access-manager
- rda-worker-all
- rda-event-gateway
- rda-bulkstats
- ubuntu-rdac
- cfxdx-nb-nginx-all
Please make sure 8.2.0.1 image tag is downloaded for the below RDAF OIA (AIOps) Application services.
- cfx rda-alert-processor
- cfx rda-collaboration
Downloaded docker images are stored under the below path.
/opt/rdaf-registry/data/docker/registry/v2/ or /opt/rdaf/data/docker/registry/v2/
Run the below command to check the filesystem's disk usage on offline registry VM where docker images are pulled.
If necessary, older image tags that are no longer in use can be deleted to free up disk space using the command below.
Note
Run the below given command if /opt occupies more than 80% of the disk space or if the free capacity of /opt is less than 25GB.
1.2.2 Upgrade RDAF Platform Services
Step-1: Run the below command to initiate upgrading RDAF Platform services.
rdafk8s platform upgrade --tag 8.2.0.1 --service rda-api-server --service rda-chat-helper --service rda-portal --service rda-collector --service rda-scheduler --service rda-resource-manager --service rda-access-manager
As the upgrade procedure is a non-disruptive upgrade, it puts the currently running PODs into Terminating state and newer version PODs into Pending state.
Step-2: Run the below command to check the status of the existing and newer PODs and make sure atleast one instance of each Platform service is in Terminating state.
Step-3: Run the below command to put all Terminating RDAF platform service PODs into maintenance mode. It will list all of the POD Ids of platform services along with rdac maintenance command that required to be put in maintenance mode.
Note
If maint_command.py script doesn't exist on RDAF deployment CLI VM, it can be downloaded using the below command.
Step-4: Copy & Paste the rdac maintenance command as below.
Step-5: Run the below command to verify the maintenance mode status of the RDAF platform services.
Step-6: Run the below command to delete the Terminating RDAF platform service PODs
for i in `kubectl get pods -n rda-fabric -l app_category=rdaf-platform | grep 'Terminating' | awk '{print $1}'`; do kubectl delete pod $i -n rda-fabric --force; done
Note
Wait for 120 seconds and Repeat above steps from Step-2 to Step-6 for rest of the RDAF Platform service PODs.
Please wait till all of the new platform service PODs are in Running state and run the below command to verify their status and make sure latest selected services are running with 8.2.0.1 version.
+--------------------+----------------+-------------------+--------------+----------+
| Name | Host | Status | Container Id | Tag |
+--------------------+----------------+-------------------+--------------+----------+
| rda-api-server | 192.168.131.44 | Up 44 Minutes ago | a1b2c3d4e5f6 | 8.2.0.1 |
| rda-api-server | 192.168.131.45 | Up 46 Minutes ago | b7c8d9e0f1a2 | 8.2.0.1 |
| rda-registry | 192.168.131.45 | Up 46 Minutes ago | c3d4e5f6a7b8 | 8.2 |
| rda-registry | 192.168.131.44 | Up 46 Minutes ago | d9e0f1a2b3c4 | 8.2 |
| rda-identity | 192.168.131.44 | Up 46 Minutes ago | e5f6a7b8c9d0 | 8.2 |
| rda-identity | 192.168.131.47 | Up 45 Minutes ago | f1a2b3c4d5e6 | 8.2 |
| rda-fsm | 192.168.131.44 | Up 46 Minutes ago | a7b8c9d0e1f2 | 8.2 |
| rda-fsm | 192.168.131.45 | Up 46 Minutes ago | b3c4d5e6f7a8 | 8.2 |
| rda-asm | 192.168.131.44 | Up 46 Minutes ago | c9d0e1f2a3b4 | 8.2 |
| rda-asm | 192.168.131.45 | Up 46 Minutes ago | d5e6f7a8b9c0 | 8.2 |
| rda-asm | 192.168.131.47 | Up 2 Weeks ago | e1f2a3b4c5d6 | 8.2 |
| rda-asm | 192.168.131.46 | Up 2 Weeks ago | f7a8b9c0d1e2 | 8.2 |
| rda-chat-helper | 192.168.131.44 | Up 46 Minutes ago | a3b4c5d6e7f8 | 8.2.0.1 |
| rda-chat-helper | 192.168.131.45 | Up 45 Minutes ago | b9c0d1e2f3a4 | 8.2.0.1 |
| rda-access-manager | 192.168.131.45 | Up 46 Minutes ago | c5d6e7f8a9b0 | 8.2.0.1 |
| rda-access-manager | 192.168.131.46 | Up 45 Minutes ago | d1e2f3a4b5c6 | 8.2.0.1 |
| rda-resource- | 192.168.131.44 | Up 45 Minutes ago | e7f8a9b0c1d2 | 8.2.0.1 |
| manager | | | | |
| rda-resource- | 192.168.131.45 | Up 45 Minutes ago | f3a4b5c6d7e8 | 8.2.0.1 |
| manager | | | | |
+--------------------+----------------+-------------------+--------------+----------+
Run the below command to check the rda-scheduler service is elected as a leader under Site column.
+-------+----------------------------------------+-------------+--------------+----------+-------------+-----------------+--------+--------------+---------------+--------------+
| Cat | Pod-Type | Pod-Ready | Host | ID | Site | Age | CPUs | Memory(GB) | Active Jobs | Total Jobs |
|-------+----------------------------------------+-------------+--------------+----------+-------------+-----------------+--------+--------------+---------------+--------------|
| Infra | api-server | True | rda-api-server | 9c0484af | | 11:41:50 | 8 | 31.33 | | |
| Infra | api-server | True | rda-api-server | 196558ed | | 11:40:23 | 8 | 31.33 | | |
| Infra | asm | True | rda-asm-5b8fb9 | bcbdaae5 | | 11:42:26 | 8 | 31.33 | | |
| Infra | asm | True | rda-asm-5b8fb9 | 232a58af | | 11:42:40 | 8 | 31.33 | | |
| Infra | collector | True | rda-collector- | d06fb56c | | 11:42:03 | 8 | 31.33 | | |
| Infra | collector | True | rda-collector- | a4c79e4c | | 11:41:59 | 8 | 31.33 | | |
| Infra | registry | True | rda-registry-6 | 2fd69950 | | 11:42:03 | 8 | 31.33 | | |
| Infra | registry | True | rda-registry-6 | fac544d6 | | 11:41:59 | 8 | 31.33 | | |
| Infra | scheduler | True | rda-scheduler- | b98afe88 | *leader* | 11:42:01 | 8 | 31.33 | | |
| Infra | scheduler | True | rda-scheduler- | e25a0841 | | 11:41:56 | 8 | 31.33 | | |
| Infra | worker | True | rda-worker-5b5 | 99bd054e | rda-site-01 | 11:33:40 | 8 | 31.33 | 0 | 0 |
| Infra | worker | True | rda-worker-5b5 | 0bfdcd98 | rda-site-01 | 11:33:34 | 8 | 31.33 | 0 | 0 |
+-------+----------------------------------------+-------------+----------------+----------+-------------+----------+--------+--------------+---------------+--------------+
Run the below command to check if all services has ok status and does not throw any failure messages.
Warning
For Non-Kubernetes deployment, upgrading RDAF Platform and AIOps application services is a disruptive operation when rolling-upgrade option is not used. Please schedule a maintenance window before upgrading RDAF Platform and AIOps services to newer version.
Run the below command to initiate upgrading RDAF Platform services with zero downtime
rdaf platform upgrade --tag 8.2.0.1 --rolling-upgrade --timeout 10 --service rda_api_server --service rda_collector --service rda_chat_helper --service portal-backend --service portal-frontend --service rda_scheduler --service cfx-rda-resource-manager --service cfx-rda-access-manager
Note
timeout <10> mentioned in the above command represents as Seconds
Note
The rolling-upgrade option upgrades the Platform services running in high-availability mode on one VM at a time in sequence. It completes the upgrade of Platform services running on VM-1 before upgrading them on VM-2, followed by VM-3, and so on.
During this upgrade sequence, RDAF platform continues to function without any impact to the application traffic.
After completing the Platform services upgrade on all VMs, it will ask for user confirmation to delete the older version Platform service PODs. The user has to provide YES to delete the old docker containers (in non-k8s)
2026-03-09 09:06:39,677 [rdaf.component.platform] INFO - Gathering platform container details.
2026-03-09 09:06:40,218 [rdaf.component.platform] INFO - Gathering rdac pod details.
+----------+-------------+---------+------------------+--------------+-------------+------------+
| Pod ID | Pod Type | Version | Age | Hostname | Maintenance | Pod Status |
+----------+-------------+---------+------------------+--------------+-------------+------------+
| 302baabf | api-server | 8.2.0.1 | 2:22:11 | c9666616012c | None | True |
| a8667566 | collector | 8.2.0.1 | 6 days, 0:47:39 | 1ae27c01b96a | None | True |
| e06533ac | chat-helper | 8.2.0.1 | 2:21:15 | 32fb76ed4cd0 | None | True |
| b5abb879 | scheduler | 8.2.0.1 | 5 days, 14:25:16 | 0725f94238b1 | None | True |
+----------+-------------+---------+------------------+--------------+-------------+------------+
Continue moving above pods to maintenance mode? [yes/no]: yes
2026-03-09 09:07:10,245 [rdaf.component.platform] INFO - Initiating Maintenance Mode...
2026-03-09 09:07:39,185 [rdaf.component.platform] INFO - Following container are in maintenance mode
+----------+-------------+---------+------------------+--------------+-------------+------------+
| Pod ID | Pod Type | Version | Age | Hostname | Maintenance | Pod Status |
+----------+-------------+---------+------------------+--------------+-------------+------------+
| 302baabf | api-server | 8.2.0.1 | 2:22:54 | c9666616012c | maintenance | False |
| e06533ac | chat-helper | 8.2.0.1 | 2:21:59 | 32fb76ed4cd0 | maintenance | False |
| a8667566 | collector | 8.2.0.1 | 6 days, 0:48:23 | 1ae27c01b96a | maintenance | False |
| b5abb879 | scheduler | 8.2.0.1 | 5 days, 14:26:00 | 0725f94238b1 | maintenance | False |
+----------+-------------+---------+------------------+--------------|-------------|------------|
Run the below command to initiate upgrading RDAF Platform services without zero downtime
rdaf platform upgrade --tag 8.2.0.1 --service rda_api_server --service rda_collector --service rda_chat_helper --service portal-backend --service portal-frontend --service rda_scheduler --service cfx-rda-resource-manager --service cfx-rda-access-manager
Please wait till all of the new platform services are in Up state and run the below command to verify their status and make sure latest selected services are running with 8.2.0.1 version.
+--------------------------+----------------+-------------------------------+--------------+----------+
| Name | Host | Status | Container Id | Tag |
+--------------------------+----------------+-------------------------------+--------------+----------+
| rda_api_server | 192.168.108.51 | Up 4 hours | 2a3b4c5d6e7f | 8.2.0.1 |
| rda_api_server | 192.168.108.52 | Up 4 hours | 9g0h1i2j3k4l | 8.2.0.1 |
| rda_registry | 192.168.108.51 | Up 4 hours | 4m5n6o7p8q9r | 8.2 |
| rda_registry | 192.168.108.52 | Up 4 hours | 1s2t3u4v5w6x | 8.2 |
| rda_scheduler | 192.168.108.51 | Up 4 hours | 7y8z9a0b1c2d | 8.2.0.1 |
| rda_scheduler | 192.168.108.52 | Up 4 hours | 4e5f6g7h8i9j | 8.2.0.1 |
| rda_collector | 192.168.108.51 | Up 4 hours | 1k2l3m4n5o6p | 8.2.0.1 |
| rda_collector | 192.168.108.52 | Up 4 hours | 8q9r0s1t2u3v | 8.2.0.1 |
| rda_identity | 192.168.108.51 | Up 4 hours | 5w6x7y8z9a0b | 8.2 |
| rda_identity | 192.168.108.52 | Up 4 hours | 2c3d4e5f6g7h | 8.2 |
| rda_asm | 192.168.108.51 | Up 4 hours | 9i0j1k2l3m4n | 8.2 |
| rda_asm | 192.168.108.52 | Up 4 hours | 6o7p8q9r0s1t | 8.2 |
| rda_fsm | 192.168.108.51 | Up 4 hours | 3u4v5w6x7y8z | 8.2 |
| rda_fsm | 192.168.108.52 | Up 4 hours | 0a1b2c3d4e5f | 8.2 |
+--------------------------+----------------+-------------------------------+--------------+----------+
Run the below command to check the rda-scheduler service is elected as a leader under Site column.
Run the below command to check if all services has ok status and does not throw any failure messages.
+-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-------------------------------------------------------------+
| Cat | Pod-Type | Host | ID | Site | Health Parameter | Status | Message |
|-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-------------------------------------------------------------|
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | service-status | ok | |
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | 192.168-connectivity | ok | |
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | service-dependency:configuration-service | ok | 2 pod(s) found for configuration-service |
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | service-initialization-status | ok | |
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | kafka-connectivity | ok | Cluster=NTc1NWU1MTQxYmY3MTFlZg, Broker=1, Brokers=[1, 2, 3] |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | service-status | ok | |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | 192.168-connectivity | ok | |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | service-dependency:configuration-service | ok | 2 pod(s) found for configuration-service |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | service-initialization-status | ok | |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | kafka-connectivity | ok | Cluster=NTc1NWU1MTQxYmY3MTFlZg, Broker=3, Brokers=[1, 2, 3] |
| rda_app | alert-processor | c6cc7b04ab33 | b4ebfb06 | | service-status | ok | |
| rda_app | alert-processor | c6cc7b04ab33 | b4ebfb06 | | 192.168-connectivity | ok | |
+-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-------------------------------------------------------------+
1.2.3 Upgrade rdac CLI
1.2.4 Upgrade RDA Worker Services
Note
If the worker was deployed in a HTTP proxy environment, please make sure the required HTTP proxy environment variables are added in /opt/rdaf/deployment-scripts/values.yaml file under rda_worker configuration section as shown below before upgrading RDA Worker services.
rda_worker:
terminationGracePeriodSeconds: 300
replicas: 6
sizeLimit: 1024Mi
privileged: true
resources:
requests:
memory: 100Mi
limits:
memory: 24Gi
env:
WORKER_GROUP: rda-prod-01
CAPACITY_FILTER: cpu_load1 <= 7.0 and mem_percent < 95
MAX_PROCESSES: '1000'
RDA_ENABLE_TRACES: 'no'
WORKER_PUBLIC_ACCESS: 'true'
DISABLE_REMOTE_LOGGING_CONTROL: 'no'
RDA_SELF_HEALTH_RESTART_AFTER_FAILURES: 3
extraEnvs:
- name: http_proxy
value: http://test:[email protected]:3128
- name: https_proxy
value: http://test:[email protected]:3128
- name: HTTP_PROXY
value: http://test:[email protected]:3128
- name: HTTPS_PROXY
value: http://test:[email protected]:3128
....
....
Step-1: Please run the below command to initiate upgrading the RDA Worker service PODs.
Step-2: Run the below command to check the status of the existing and newer PODs and make sure atleast one instance of each RDA Worker service POD is in Terminating state.
NAME READY STATUS RESTARTS AGE
rda-worker-77f459d5b9-9kdmg 1/1 Running 0 73m
rda-worker-77f459d5b9-htsmr 1/1 Running 0 74m
Step-3: Run the below command to put all Terminating RDAF worker service PODs into maintenance mode. It will list all of the POD Ids of RDA worker services along with rdac maintenance command that is required to be put in maintenance mode.
Step-4: Copy & Paste the rdac maintenance command as below.
Step-5: Run the below command to verify the maintenance mode status of the RDAF worker services.
Step-6: Run the below command to delete the Terminating RDAF worker service PODs
for i in `kubectl get pods -n rda-fabric -l app_component=rda-worker | grep 'Terminating' | awk '{print $1}'`; do kubectl delete pod $i -n rda-fabric --force; done
Note
Wait for 120 seconds between each RDAF worker service upgrade by repeating above steps from Step-2 to Step-6 for rest of the RDAF worker service PODs.
Step-7: Please wait for 120 seconds to let the newer version of RDA Worker service PODs join the RDA Fabric appropriately. Run the below commands to verify the status of the newer RDA Worker service PODs.
+------------+----------------+---------------+--------------+----------+
| Name | Host | Status | Container Id | Tag |
+------------+----------------+---------------+--------------+----------+
| rda-worker | 192.168.108.17 | Up 1 Hour ago | a1b2c3d4e5f0 | 8.2.0.1 |
| rda-worker | 192.168.108.18 | Up 1 Hour ago | f9e8d7c6b5a4 | 8.2.0.1 |
+------------+----------------+---------------+--------------+----------+
Step-8: Run the below command to check if all RDA Worker services has ok status and does not throw any failure messages.
Note
If the worker was deployed in a HTTP proxy environment, please make sure the required HTTP proxy environment variables are added in /opt/rdaf/deployment-scripts/values.yaml file under rda_worker configuration section as shown below before upgrading RDA Worker services.
rda_worker:
mem_limit: 8G
memswap_limit: 8G
privileged: false
environment:
RDA_ENABLE_TRACES: 'no'
RDA_SELF_HEALTH_RESTART_AFTER_FAILURES: 3
http_proxy: "http://test:[email protected]:3128"
https_proxy: "http://test:[email protected]:3128"
HTTP_PROXY: "http://test:[email protected]:3128"
HTTPS_PROXY: "http://test:[email protected]:3128"
- Upgrade RDA Worker Services
Please run the below command to initiate upgrading the RDA Worker Service with zero downtime
Note
timeout <10> mentioned in the above command represents as seconds
Note
The rolling-upgrade option upgrades the Worker services running in high-availability mode on one VM at a time in sequence. It completes the upgrade of Worker services running on VM-1 before upgrading them on VM-2, followed by VM-3, and so on.
After completing the Worker services upgrade on all VMs, it will ask for user confirmation, the user has to provide YES to delete the older version Worker service PODs.
2024-08-12 02:56:11,573 [rdaf.component.worker] INFO - Collecting worker details for rolling upgrade
2024-08-12 02:56:14,301 [rdaf.component.worker] INFO - Rolling upgrade worker on 192.168.133.96
+----------+----------+---------------+---------+--------------+-------------+------------+
| Pod ID | Pod Type | Version | Age | Hostname | Maintenance | Pod Status |
+----------+----------+---------------+---------+--------------+-------------+------------+
| c8a37db9 | worker | 8.2.0.1 |3:32:31 | fffe44b43708 | None | True |
+----------+----------+---------------+---------+--------------+-------------+------------+
Continue moving above pod to maintenance mode? [yes/no]: yes
2024-08-12 02:57:17,346 [rdaf.component.worker] INFO - Initiating maintenance mode for pod c8a37db9
2024-08-12 02:57:22,401 [rdaf.component.worker] INFO - Waiting for worker to be moved to maintenance.
2024-08-12 02:57:35,001 [rdaf.component.worker] INFO - Following worker container is in maintenance mode
+----------+----------+---------------+---------+--------------+-------------+------------+
| Pod ID | Pod Type | Version | Age | Hostname | Maintenance | Pod Status |
+----------+----------+---------------+---------+--------------+-------------+------------+
| c8a37db9 | worker | 8.2.0.1 | 3:33:52 | fffe44b43708 | maintenance | False |
+----------+----------+---------------+---------+--------------+-------------+------------+
2024-08-12 02:57:35,002 [rdaf.component.worker] INFO - Waiting for timeout of 3 seconds.
Please run the below command to initiate upgrading the RDA Worker Service without zero downtime
Please wait for 120 seconds to let the newer version of RDA Worker service containers join the RDA Fabric appropriately. Run the below commands to verify the status of the newer RDA Worker service containers.
| Infra | worker | True | 6eff605e72c4 | a318f394 | rda-site-01 | 13:45:13 | 4 | 31.21 | 0 | 0 |
| Infra | worker | True | ae7244d0d10a | 554c2cd8 | rda-site-01 | 13:40:40 | 4 | 31.21 | 0 | 0 |
+------------+----------------+------------+--------------+---------+
| Name | Host | Status | Container Id | Tag |
+------------+----------------+------------+--------------+---------+
| rda_worker | 192.168.108.53 | Up 4 hours | 1a2b3c4d5e6f | 8.2.0.1 |
| rda_worker | 192.168.108.54 | Up 4 hours | 9g8h7i6j5k4l | 8.2.0.1 |
+------------+----------------+------------+--------------+---------+
+-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------------------------+
| Cat | Pod-Type | Host | ID | Site | Health Parameter | Status | Message |
|-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------------------------|
| rda_infra | api-server | 1b0542719618 | 1845ae67 | | service-status | ok | |
| rda_infra | api-server | 1b0542719618 | 1845ae67 | | 192.168-connectivity | ok | |
| rda_infra | api-server | d4404cffdc7a | a4cfdc6d | | service-status | ok | |
| rda_infra | api-server | d4404cffdc7a | a4cfdc6d | | 192.168-connectivity | ok | |
| rda_infra | asm | 8d3d52a7a475 | 418c9dc1 | | service-status | ok | |
| rda_infra | asm | 8d3d52a7a475 | 418c9dc1 | | 192.168-connectivity | ok | |
| rda_infra | asm | ab172a9b8229 | 2ac1d67a | | service-status | ok | |
| rda_infra | asm | ab172a9b8229 | 2ac1d67a | | 192.168-connectivity | ok | |
| rda_app | asset-dependency | 6ac69ca1085c | c2e9dcb9 | | service-status | ok | |
| rda_app | asset-dependency | 6ac69ca1085c | c2e9dcb9 | | 192.168-connectivity | ok | |
| rda_app | asset-dependency | 58a5f4f460d3 | 0b91caac | | service-status | ok | |
| rda_app | asset-dependency | 58a5f4f460d3 | 0b91caac | | 192.168-connectivity | ok | |
| rda_app | authenticator | 9011c2aef498 | 9f7efdc3 | | service-status | ok | |
| rda_app | authenticator | 9011c2aef498 | 9f7efdc3 | | 192.168-connectivity | ok | |
| rda_app | authenticator | 9011c2aef498 | 9f7efdc3 | | DB-connectivity | ok | |
| rda_app | authenticator | 148621ed8c82 | dbf16b82 | | service-status | ok | |
| rda_app | authenticator | 148621ed8c82 | dbf16b82 | | 192.168-connectivity | ok | |
| rda_app | authenticator | 148621ed8c82 | dbf16b82 | | DB-connectivity | ok | |
| rda_app | cfx-app-controller | 75ec0f30cfa3 | 1198fdee | | service-status | ok | |
| rda_app | cfx-app-controller | 75ec0f30cfa3 | 1198fdee | | 192.168-connectivity | ok | |
| rda_app | cfx-app-controller | 75ec0f30cfa3 | 1198fdee | | service-initialization-status | ok | |
| rda_app | cfx-app-controller | 75ec0f30cfa3 | 1198fdee | | DB-connectivity | ok |
+-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------------------------+
1.2.5 Upgrade OIA Application Services
Step-1: Run the below commands to initiate upgrading RDAF OIA Application services
rdafk8s app upgrade OIA --tag 8.2.0.1 --service rda-alert-processor --service rda-alert-correlator --service rda-collaboration
Step-2: Run the below command to check the status of the newly upgraded PODs.
Step-3: Run the below command to put all Terminating OIA application service PODs into maintenance mode. It will list all of the POD Ids of OIA application services along with rdac maintenance command that are required to be put in maintenance mode.
Step-4: Copy & Paste the rdac maintenance command as below.
Step-5: Run the below command to verify the maintenance mode status of the OIA application services.
Step-6: Run the below command to delete the Terminating OIA application service PODs
for i in `kubectl get pods -n rda-fabric -l app_name=oia | grep 'Terminating' | awk '{print $1}'`; do kubectl delete pod $i -n rda-fabric --force; done
Note
Wait for 120 seconds and Repeat above steps from Step-2 to Step-6 for rest of the OIA application service PODs.
Please wait till the selected OIA application service PODs are in Running state and run the below command to verify their status and make sure they are running with 8.2 / 8.2.0.1 version.
+--------------------+----------------+-------------------+--------------+---------+
| Name | Host | Status | Container Id | Tag |
+--------------------+----------------+-------------------+--------------+---------+
| rda-alert-ingester | 192.168.131.47 | Up 54 Minutes ago | a1b2c3d4e5f6 | 8.2 |
| rda-alert-ingester | 192.168.131.49 | Up 49 Minutes ago | b7c8d9e0f1a2 | 8.2 |
| rda-alert- | 192.168.131.49 | Up 44 Minutes ago | c3d4e5f6a7b8 | 8.2.0.1 |
| processor | | | | |
| rda-alert- | 192.168.131.50 | Up 54 Minutes ago | d9e0f1a2b3c4 | 8.2.0.1 |
| processor | | | | |
| rda-alert- | 192.168.131.47 | Up 54 Minutes ago | e5f6a7b8c9d0 | 8.2 |
| processor- | | | | |
| companion | | | | |
| rda-alert- | 192.168.131.49 | Up 48 Minutes ago | f1a2b3c4d5e6 | 8.2 |
| processor- | | | | |
| companion | | | | |
| rda-app-controller | 192.168.131.47 | Up 54 Minutes ago | a7b8c9d0e1f2 | 8.2 |
| rda-app-controller | 192.168.131.46 | Up 54 Minutes ago | b3c4d5e6f7a8 | 8.2 |
| rda-collaboration | 192.168.131.49 | Up 43 Minutes ago | c9d0e1f2a3b4 | 8.2.0.1 |
| rda-collaboration | 192.168.131.50 | Up 53 Minutes ago | d5e6f7a8b9c0 | 8.2.0.1 |
| rda-configuration- | 192.168.131.46 | Up 54 Minutes ago | e1f2a3b4c5d6 | 8.2 |
| service | | | | |
| rda-configuration- | 192.168.131.49 | Up 51 Minutes ago | f7a8b9c0d1e2 | 8.2 |
| service | | | | |
+--------------------+----------------+-------------------+--------------+---------+
Step-7: Run the below command to verify all OIA application services are up and running.
+-------+----------------------------------------+-------------+----------------+----------+-------------+----------+--------+--------------+---------------+--------------+
| Cat | Pod-Type | Pod-Ready | Host | ID | Site | Age | CPUs | Memory(GB) | Active Jobs | Total Jobs |
|-------+----------------------------------------+-------------+----------------+----------+-------------+----------+--------+--------------+---------------+--------------|
| App | alert-ingester | True | rda-alert-inge | 6a6e464d | | 19:19:06 | 8 | 31.33 | | |
| App | alert-ingester | True | rda-alert-inge | 7f6b42a0 | | 19:19:23 | 8 | 31.33 | | |
| App | alert-processor | True | rda-alert-proc | a880e491 | | 19:19:51 | 8 | 31.33 | | |
| App | alert-processor | True | rda-alert-proc | b684609e | | 19:19:48 | 8 | 31.33 | | |
| App | alert-processor-companion | True | rda-alert-proc | 874f3b33 | | 19:18:54 | 8 | 31.33 | | |
| App | alert-processor-companion | True | rda-alert-proc | 70cadaa7 | | 19:18:35 | 8 | 31.33 | | |
| App | asset-dependency | True | rda-asset-depe | bde06c15 | | 19:44:20 | 8 | 31.33 | | |
| App | asset-dependency | True | rda-asset-depe | 47b9eb02 | | 19:44:08 | 8 | 31.33 | | |
| App | authenticator | True | rda-identity-d | faa33e1b | | 19:44:22 | 8 | 31.33 | | |
| App | authenticator | True | rda-identity-d | 36083c36 | | 19:44:16 | 8 | 31.33 | | |
| App | cfx-app-controller | True | rda-app-contro | 5fd3c3f4 | | 19:19:39 | 8 | 31.33 | | |
| App | cfx-app-controller | True | rda-app-contro | d66e5ce8 | | 19:19:26 | 8 | 31.33 | | |
| App | cfxdimensions-app-access-manager | True | rda-access-man | ecbb535c | | 19:44:16 | 8 | 31.33 | | |
| App | cfxdimensions-app-access-manager | True | rda-access-man | 9a05db5a | | 19:44:06 | 8 | 31.33 | | |
| App | cfxdimensions-app-collaboration | True | rda-collaborat | 61b3c53b | | 19:18:48 | 8 | 31.33 | | |
| App | cfxdimensions-app-collaboration | True | rda-collaborat | 09b9474e | | 19:18:27 | 8 | 31.33 | | |
+-------+----------------------------------------+-------------+----------------+----------+-------------+-------------------+--------+-----------------------------+--------------+
Run the below command to check if all services has ok status and does not throw any failure messages.
+-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------------------------+
| Cat | Pod-Type | Host | ID | Site | Health Parameter | Status | Message |
|-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-----------------------------------------------------------------------------------------------------------------------------|
| rda_app | alert-ingester | rda-alert-in | 6a6e464d | | service-status | ok | |
| rda_app | alert-ingester | rda-alert-in | 6a6e464d | | 192.168-connectivity | ok | |
| rda_app | alert-ingester | rda-alert-in | 6a6e464d | | service-dependency:configuration-service | ok | 2 pod(s) found for configuration-service |
| rda_app | alert-ingester | rda-alert-in | 6a6e464d | | service-initialization-status | ok | |
| rda_app | alert-ingester | rda-alert-in | 6a6e464d | | kafka-connectivity | ok | Cluster=dKnnkaYSPELK8DBUk0rPig, Broker=0, Brokers=[0, 1, 2] |
| rda_app | alert-ingester | rda-alert-in | 6a6e464d | | kafka-consumer | ok | Health: [{'387c0cb507b84878b9d0b15222cb4226.inbound-events': 0, '387c0cb507b84878b9d0b15222cb4226.mapped-events': 0}, {}] |
| rda_app | alert-ingester | rda-alert-in | 7f6b42a0 | | service-status | ok | |
| rda_app | alert-ingester | rda-alert-in | 7f6b42a0 | | 192.168-connectivity | ok | |
| rda_app | alert-ingester | rda-alert-in | 7f6b42a0 | | service-dependency:configuration-service | ok | 2 pod(s) found for configuration-service |
| rda_app | alert-ingester | rda-alert-in | 7f6b42a0 | | service-initialization-status | ok | |
| rda_app | alert-ingester | rda-alert-in | 7f6b42a0 | | kafka-consumer | ok | Health: [{'387c0cb507b84878b9d0b15222cb4226.inbound-events': 0, '387c0cb507b84878b9d0b15222cb4226.mapped-events': 0}, {}] |
| rda_app | alert-ingester | rda-alert-in | 7f6b42a0 | | kafka-connectivity | ok | Cluster=dKnnkaYSPELK8DBUk0rPig, Broker=1, Brokers=[0, 1, 2] |
| rda_app | alert-processor | rda-alert-pr | a880e491 | | service-status | ok | |
| rda_app | alert-processor | rda-alert-pr | a880e491 | | 192.168-connectivity | ok | |
| rda_app | alert-processor | rda-alert-pr | a880e491 | | service-dependency:cfx-app-controller | ok | 2 pod(s) found for cfx-app-controller |
| rda_app | alert-processor | rda-alert-pr | a880e491 | | service-dependency:configuration-service | ok | 2 pod(s) found for configuration-service |
| rda_app | alert-processor | rda-alert-pr | a880e491 | | service-initialization-status | ok | |
| rda_app | alert-processor | rda-alert-pr | a880e491 | | kafka-connectivity | ok | Cluster=dKnnkaYSPELK8DBUk0rPig, Broker=1, Brokers=[0, 1, 2] |
| rda_app | alert-processor | rda-alert-pr | a880e491 | | DB-connectivity | ok | |
+-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-------------------------------------------------------------+
Run the below commands to initiate upgrading the RDA Fabric OIA Application services with zero downtime
rdaf app upgrade OIA --tag 8.2.0.1 --rolling-upgrade --timeout 10 --service rda-alert-processor --service rda-alert-correlator --service rda-collaboration
Note
timeout <10> mentioned in the above command represents as Seconds
Note
The rolling-upgrade option upgrades the OIA application services running in high-availability mode on one VM at a time in sequence. It completes the upgrade of OIA application services running on VM-1 before upgrading them on VM-2, followed by VM-3, and so on.
After completing the OIA application services upgrade on all VMs, it will ask for user confirmation to delete the older version OIA application service PODs.
2024-08-12 03:18:08,705 [rdaf.component.oia] INFO - Gathering OIA app container details.
2024-08-12 03:18:10,719 [rdaf.component.oia] INFO - Gathering rdac pod details.
+----------+----------------------+---------+---------+--------------+-------------+------------+
| Pod ID | Pod Type | Version | Age | Hostname | Maintenance | Pod Status |
+----------+----------------------+---------+---------+--------------+-------------+------------+
| 2992fe69 | cfx-app-controller | 8.2 | 3:44:53 | a1b2c3d4e5f0 | None | True |
| 336138c8 | reports-registry | 8.2 | 3:44:12 | b7c8d9e0f1a2 | None | True |
| ccc5f3ce | cfxdimensions-app- | 8.2 | 3:43:34 | c3d4e5f6a7b8 | None | True |
| | notification-service | 8.2 | | | | |
| 03614007 | cfxdimensions-app- | 8.2 | 3:42:54 | d9e0f1a2b3c4 | None | True |
| | file-browser | 8.2 | | | | |
| a4949804 | configuration- | 8.2 | 3:42:15 | e5f6a7b8c9d0 | None | True |
| | service | 8.2 | | | | |
| 8f37c520 | alert-ingester | 8.2 | 3:41:35 | f1a2b3c4d5e6 | None | True |
| 249b7104 | webhook-server | 8.2 | 3:12:04 | a7b8c9d0e1f2 | None | True |
| 76c64336 | smtp-server | 8.2 | 3:08:57 | b3c4d5e6f7a8 | None | True |
| ad85cb4c | event-consumer | 8.2 | 3:09:58 | c9d0e1f2a3b4 | None | True |
| 1a788ef3 | alert-processor | 8.2 | 3:11:01 | d5e6f7a8b9c0 | None | True |
| 970b90b1 | cfxdimensions-app- | 8.2 | 3:38:14 | e1f2a3b4c5d6 | None | True |
| | irm_service | 8.2 | | | | |
| 153aa6ac | ml-config | 8.2 | 3:37:33 | f7a8b9c0d1e2 | None | True |
| 5aa927a4 | cfxdimensions-app- | 8.2 | 3:36:53 | a3b4c5d6e7f8 | None | True |
| | collaboration | 8.2 | | | | |
| 6833aa86 | ingestion-tracker | 8.2 | 3:36:13 | b9c0d1e2f3a4 | None | True |
| afe77cb9 | alert-processor- | 8.2 | 3:35:33 | c5d6e7f8a9b0 | None | True |
| | companion | 8.2 | | | | |
+----------+----------------------+---------+---------+--------------+-------------+------------+
Continue moving above pods to maintenance mode? [yes/no]: yes
2024-08-12 03:18:27,159 [rdaf.component.oia] INFO - Initiating Maintenance Mode...
2024-08-12 03:18:32,978 [rdaf.component.oia] INFO - Waiting for services to be moved to maintenance.
2024-08-12 03:18:55,771 [rdaf.component.oia] INFO - Following container are in maintenance mode
+----------+----------------------+---------+---------+--------------+-------------+------------+
Run the below command to initiate upgrading the RDA Fabric OIA Application services without zero downtime
rdaf app upgrade OIA --tag 8.2.0.1 --service rda-alert-processor --service rda-alert-correlator --service rda-collaboration
Please wait till the selected OIA application service containers are in Up state and run the below command to verify their status and make sure they are running with 8.2 / 8.2.0.1 version.
+-----------------------------------+----------------+------------+--------------+---------+
| Name | Host | Status | Container Id | Tag |
+-----------------------------------+----------------+------------+--------------+---------+
| cfx-rda-app-controller | 192.168.108.51 | Up 3 hours | a1b2c3d4e5f0 | 8.2 |
| cfx-rda-app-controller | 192.168.108.52 | Up 3 hours | f9e8d7c6b5a4 | 8.2 |
| cfx-rda-reports-registry | 192.168.108.51 | Up 4 hours | c7d8e9f0a1b2 | 8.2 |
| cfx-rda-reports-registry | 192.168.108.52 | Up 4 hours | a3b4c5d6e7f8 | 8.2 |
| cfx-rda-notification-service | 192.168.108.51 | Up 4 hours | b9c0d1e2f3a4 | 8.2 |
| cfx-rda-notification-service | 192.168.108.52 | Up 4 hours | c5d6e7f8a9b0 | 8.2 |
| cfx-rda-file-browser | 192.168.108.51 | Up 4 hours | d1e2f3a4b5c6 | 8.2 |
| cfx-rda-file-browser | 192.168.108.52 | Up 4 hours | e7f8a9b0c1d2 | 8.2 |
| cfx-rda-configuration-service | 192.168.108.51 | Up 4 hours | f3a4b5c6d7e8 | 8.2 |
| cfx-rda-configuration-service | 192.168.108.52 | Up 4 hours | a9b0c1d2e3f4 | 8.2 |
| cfx-rda-alert-ingester | 192.168.108.51 | Up 4 hours | b5c6d7e8f9a0 | 8.2 |
| cfx-rda-alert-ingester | 192.168.108.52 | Up 4 hours | c1d2e3f4a5b6 | 8.2 |
| cfx-rda-webhook-server | 192.168.108.51 | Up 4 hours | d7e8f9a0b1c2 | 8.2 |
| cfx-rda-webhook-server | 192.168.108.52 | Up 4 hours | e3f4a5b6c7d8 | 8.2 |
| cfx-rda-smtp-server | 192.168.108.51 | Up 4 hours | f9a0b1c2d3e4 | 8.2 |
| cfx-rda-smtp-server | 192.168.108.52 | Up 4 hours | a5b6c7d8e9f0 | 8.2 |
+-----------------------------------+----------------+------------+--------------+---------+
Run the below command to verify all OIA application services are up and running.
+-------+----------------------------------------+-------------+----------------+----------+-------------+----------+--------+--------------+---------------+--------------+
| Cat | Pod-Type | Pod-Ready | Host | ID | Site | Age | CPUs | Memory(GB) | Active Jobs | Total Jobs |
|-------+----------------------------------------+-------------+----------------+----------+-------------+----------+--------+--------------+---------------+--------------|
| App | alert-ingester | True | rda-alert-inge | 6a6e464d | | 19:22:36 | 8 | 31.33 | | |
| App | alert-ingester | True | rda-alert-inge | 7f6b42a0 | | 19:22:53 | 8 | 31.33 | | |
| App | alert-processor | True | rda-alert-proc | a880e491 | | 19:23:21 | 8 | 31.33 | | |
| App | alert-processor | True | rda-alert-proc | b684609e | | 19:23:18 | 8 | 31.33 | | |
| App | alert-processor-companion | True | rda-alert-proc | 874f3b33 | | 19:22:24 | 8 | 31.33 | | |
| App | alert-processor-companion | True | rda-alert-proc | 70cadaa7 | | 19:22:05 | 8 | 31.33 | | |
| App | asset-dependency | True | rda-asset-depe | bde06c15 | | 19:47:50 | 8 | 31.33 | | |
| App | asset-dependency | True | rda-asset-depe | 47b9eb02 | | 19:47:38 | 8 | 31.33 | | |
| App | authenticator | True | rda-identity-d | faa33e1b | | 19:47:52 | 8 | 31.33 | | |
| App | authenticator | True | rda-identity-d | 36083c36 | | 19:47:46 | 8 | 31.33 | | |
| App | cfx-app-controller | True | rda-app-contro | 5fd3c3f4 | | 19:23:09 | 8 | 31.33 | | |
| App | cfx-app-controller | True | rda-app-contro | d66e5ce8 | | 19:22:56 | 8 | 31.33 | | |
| App | cfxdimensions-app-access-manager | True | rda-access-man | ecbb535c | | 19:47:46 | 8 | 31.33 | | |
| App | cfxdimensions-app-access-manager | True | rda-access-man | 9a05db5a | | 19:47:36 | 8 | 31.33 | | |
| App | cfxdimensions-app-collaboration | True | rda-collaborat | 61b3c53b | | 19:22:18 | 8 | 31.33 | | |
| App | cfxdimensions-app-collaboration | True | rda-collaborat | 09b9474e | | 19:21:57 | 8 | 31.33 | | |
| App | cfxdimensions-app-file-browser | True | rda-file-brows | 00495640 | | 19:22:45 | 8 | 31.33 | | |
| App | cfxdimensions-app-file-browser | True | rda-file-brows | 640f0653 | | 19:22:29 | 8 | 31.33 | | |
| App | cfxdimensions-app-irm_service | True | rda-irm-servic | 27e345c5 | | 19:21:43 | 8 | 31.33 | | |
| App | cfxdimensions-app-irm_service | True | rda-irm-servic | 23c7e082 | | 19:21:56 | 8 | 31.33 | | |
| App | cfxdimensions-app-notification-service | True | rda-notificati | bbb5b08b | | 19:23:20 | 8 | 31.33 | | |
| App | cfxdimensions-app-notification-service | True | rda-notificati | 9841bcb5 | | 19:23:02 | 8 | 31.33 | | |
+-------+----------------------------------------+-------------+----------------+----------+-------------+----------+--------+--------------+---------------+--------------+
Run the below command to check if all services has ok status and does not throw any failure messages.
+-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-------------------------------------------------------------+
| Cat | Pod-Type | Host | ID | Site | Health Parameter | Status | Message |
|-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-------------------------------------------------------------|
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | service-status | ok | |
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | 192.168-connectivity | ok | |
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | service-dependency:configuration-service | ok | 2 pod(s) found for configuration-service |
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | service-initialization-status | ok | |
| rda_app | alert-ingester | 7f75047e9e44 | daa8c414 | | kafka-connectivity | ok | Cluster=NTc1NWU1MTQxYmY3MTFlZg, Broker=1, Brokers=[1, 2, 3] |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | service-status | ok | |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | 192.168-connectivity | ok | |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | service-dependency:configuration-service | ok | 2 pod(s) found for configuration-service |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | service-initialization-status | ok | |
| rda_app | alert-ingester | f9ec55862be0 | f9b9231c | | kafka-connectivity | ok | Cluster=NTc1NWU1MTQxYmY3MTFlZg, Broker=2, Brokers=[1, 2, 3] |
| rda_app | alert-processor | c6cc7b04ab33 | b4ebfb06 | | service-status | ok | |
| rda_app | alert-processor | c6cc7b04ab33 | b4ebfb06 | | 192.168-connectivity | ok | |
+-----------+----------------------------------------+--------------+----------+-------------+-----------------------------------------------------+----------+-------------------------------------------------------------+
1.2.6 Upgrade Event Gateway Services
Step 1. Prerequisites
- Event Gateway with 8.2 tag should be already installed
Step 2. Upgrade Event Gateway Using RDAF CLI
- To upgrade the event gateway, log in to the rdaf cli VM and execute the following command.
+-------------------+-----------------+---------------+--------------+-------------+
| Name | Host | Status | Container Id | Tag |
+-------------------+-----------------+---------------+--------------+-------------+
| rda-event-gateway | 192.168.108.118 | Up 1 Days ago | 75e8baae6bbc | 8.2.0.1 |
| rda-event-gateway | 192.168.108.117 | Up 1 Days ago | 53ea97a898c0 | 8.2.0.1 |
+-------------------+-----------------+---------------+--------------+-------------+
Step 1. Prerequisites
- Event Gateway with 8.2 tag should be already installed
Step 2. Upgrade Event Gateway Using RDAF CLI
- To upgrade the event gateway, log in to the rdaf cli VM and execute the following command.
+-------------------+-----------------+---------------+--------------+---------+
| Name | Host | Status | Container Id | Tag |
+-------------------+-----------------+---------------+--------------+---------+
| rda_event_gateway | 192.168.108.127 | Up 42 hours | c22b1cf6900e | 8.2.0.1 |
| rda_event_gateway | 192.168.108.128 | Up 42 hours | 36b86a7bdff3 | 8.2.0.1 |
+-------------------+-----------------+---------------+--------------+---------+
1.2.7 Upgrade RDAF Bulkstats Services
Note
This service is applicable for Non-K8s only
Note
The RDAF Bulkstats service is optional and only necessary if the Bulkstats data ingestion feature is required. Otherwise, you may ignore the steps below and go to next section.
Run the below command to upgrade bulk_stats services
Run the below command to get the bulk_stats status
+----------------+----------------+------------+--------------+---------+
| Name | Host | Status | Container Id | Tag |
+----------------+----------------+------------+--------------+---------+
| rda_bulk_stats | 192.168.108.51 | Up 4 hours | c7d8e9f0a1b2 | 8.2.0.1 |
| rda_bulk_stats | 192.168.108.52 | Up 4 hours | a3b4c5d6e7f8 | 8.2.0.1 |
+----------------+----------------+------------+--------------+---------+
1.2.8 RDA Studio Upgrade
Please navigate to the rda-studio.yml file. You need to modify the existing tag version to 8.2.0.1, ensuring it matches the format shown in the example below, and then save the file
services:
cfxdx:
image: docker2.cloudfabrix.io:443/external/ubuntu-cfxdx-nb-nginx-all:8.2.0.1
restart: unless-stopped
volumes:
- /opt/rdaf/cfxdx/home/:/root
- /opt/rdaf/cfxdx/config/:/tmp/config/
- /opt/rdaf/cfxdx/output:/tmp/output/
- /opt/rdaf/config/network_config/:/network_config
ports:
- "9998:9998"
environment:
#JUPYTER_TOKEN: cfxdxdemo
NLTK_DATA : "/root/nltk_data"
CFXDX_CONFIG_FILE: /tmp/config/conf.yml
RDA_NETWORK_CONFIG: /network_config/config.json
RDA_USER: xxxxxxx
RDA_PASSWORD: xxxxxxxxxxxx
After updating the rda-studio.yml file to set the tag version to 8.2.0.1, execute the following commands to pull the latest images and start the services
1.2.9 Prune Images
After upgrading the services, run the below command to clean up the un-used docker images. This command helps to clean up and free the disk space
1.3. Post Upgrade Steps
Note
To get the latest OIA Alerts and Incidents Dashboard changes, please activate the Fabrix AIOps Fault Management Base Version 11.0.0
1) Upload following RDA Packs (Go to Main Menu --> Configuration --> RDA Administration --> Packs --> Click Upload Packs from Catalog), and activate the packs for the latest dashboard changes for OIA Alerts and Incidents. Upload the following pack (ensure you select the correct versions)
- Fabrix AIOps Fault Management Base Version 11.0.0
1) Download the Fabrix AIOps Fault Management Base Version 11.0.0 from the following Link
Upload the downloaded RDA Packs Go to Main Menu --> Configuration --> RDA Administration --> Packs --> Click Upload Pack
Note
To get the latest OIA Alerts and Incidents Dashboard changes, please activate the Fabrix AIOps Fault Management Base Version 11.0.0
1) Upload following RDA Packs (Go to Main Menu --> Configuration --> RDA Administration --> Packs --> Click Upload Packs from Catalog), and activate the packs for the latest dashboard changes for OIA Alerts and Incidents. Upload the following pack (ensure you select the correct versions)
- Fabrix AIOps Fault Management Base Version 11.0.0
1) Download the Fabrix AIOps Fault Management Base Version 11.0.0 from the following Link
Upload the downloaded RDA Packs Go to Main Menu --> Configuration --> RDA Administration --> Packs --> Click Upload Pack
Note
The 2nd and 3rd steps apply only to the Incident topology. If the stack contains a large volume of data, the topology may time out and fail to load.
2) Need to update include_alert_nodes from "true" to "false"(include_alert_nodes": true) in below dashboards as shown in the below screenshot
Navigation Path : Go to Main Menu --> Configuration -> RDA Administration -> User Dashboards -> Click on Edit Json on the Respective Dashboard
a) oia-incidents-os-template
b) oia-alerts-os
c) incident-topology
d) incident-details-topology
3) Need to update set related_alerts.enabled=true to related_alerts.enabled=false in Topology page configuration
Navigation Path : Go to Main Menu --> Configuration -> Apps Administration -> Configure -> Incidents -> Incident Page Configuration -> Topology -> Edit config -> and set **related_alerts.enabled=false** as shown in the below screenshot
4) To configure the "Agentic AI" features, please click here

