Uninstalling from RHOSP
Clean Trilio Datamover API service
The following steps need to be run on all nodes, which have the Trilio Datamover API service running. Those nodes can be identified by checking the roles_data.yaml
for the role that contains the entry OS::TripleO::Services::TrilioDatamoverApi
.
Once the role that runs the Trilio Datamover API service has been identified will the following commands clean the nodes from the service.
Run all commands as root or user with sudo permissions.
Stop trilio_dmapi
container.
# For RHOSP13
systemctl disable tripleo_trilio_dmapi.service
systemctl stop tripleo_trilio_dmapi.service
docker stop trilio_dmapi
# For RHOSP16 onwards
systemctl disable tripleo_trilio_dmapi.service
systemctl stop tripleo_trilio_dmapi.service
podman stop trilio_dmapi
Remove trilio_dmapi
container.
# For RHOSP13
docker rm trilio_dmapi
docker rm trilio_datamover_api_init_log
docker rm trilio_datamover_api_db_sync
# For RHOSP16 onwards
podman rm trilio_dmapi
podman rm trilio_datamover_api_init_log
podman rm trilio_datamover_api_db_sync
## If present, remove below container as well
podman rm container-puppet-triliodmapi
Clean Trilio Datamover API service conf directory.
rm -rf /var/lib/config-data/puppet-generated/triliodmapi
rm /var/lib/config-data/puppet-generated/triliodmapi.md5sum
rm -rf /var/lib/config-data/triliodmapi*
Clean Trilio Datamover API service log directory.
rm -rf /var/log/containers/trilio-datamover-api/
Clean Trilio Datamover Service
The following steps need to be run on all nodes, which have the Trilio Datamover service running. Those nodes can be identified by checking the roles_data.yaml
for the role that contains the entry OS::TripleO::Services::TrilioDatamover
.
Once the role that runs the Trilio Datamover service has been identified will the following commands clean the nodes from the service.
Run all commands as root or user with sudo permissions.
Stop trilio_datamover
container.
# For RHOSP13
docker stop trilio_datamover
# For RHOSP16 onwards
systemctl disable tripleo_trilio_datamover.service
systemctl stop tripleo_trilio_datamover.service
podman stop trilio_datamover
Remove trilio_datamover
container.
# For RHOSP13
docker rm trilio_datamover
# For RHOSP16 onwards
podman rm trilio_datamover
## If present, remove below container as well
podman rm container-puppet-triliodmapi
Unmount Trilio Backup Target on compute host.
## Following steps applicable for all supported RHOSP releases.
# Check triliovault backup target mount point
mount | grep trilio
# Unmount it
-- If it's NFS (COPY UUID_DIR from your compute host using above command)
umount /var/lib/nova/triliovault-mounts/<UUID_DIR>
-- If it's S3
umount /var/lib/nova/triliovault-mounts
# Verify that it's unmounted
mount | grep trilio
df -h | grep trilio
# Remove mount point directory after verifying that backup target unmounted successfully.
# Otherwise actual data from backup target may get cleaned.
rm -rf /var/lib/nova/triliovault-mounts
Clean Trilio Datamover service conf directory.
rm -rf /var/lib/config-data/puppet-generated/triliodm/
rm /var/lib/config-data/puppet-generated/triliodm.md5sum
rm -rf /var/lib/config-data/triliodm*
Clean log directory of Trilio Datamover service.
rm -rf /var/log/containers/trilio-datamover/
Clean Trilio haproxy resources
The following steps need to be run on all nodes, which have the haproxy service running. Those nodes can be identified by checking the roles_data.yaml
for the role that contains the entry OS::TripleO::Services::HAproxy
.
Once the role that runs the haproxy service has been identified will the following commands clean the nodes from all Trilio resources.
Run all commands as root or user with sudo permissions.
Edit the following file inside the haproxy container and remove all Trilio entries.
/var/lib/config-data/puppet-generated/haproxy/etc/haproxy/haproxy.cfg
An example of these entries is given below.
listen trilio_datamover_api
bind 172.25.3.60:13784 transparent ssl crt /etc/pki/tls/private/overcloud_endpoint.pem
bind 172.25.3.60:8784 transparent
http-request set-header X-Forwarded-Proto https if { ssl_fc }
http-request set-header X-Forwarded-Proto http if !{ ssl_fc }
http-request set-header X-Forwarded-Port %[dst_port]
option httpchk
option httplog
server overcloud-controller-0.internalapi.localdomain 172.25.3.59:8784 check fall 5 inter 2000 rise 2
Restart the haproxy container once all edits have been done.
# For RHOSP13
docker restart haproxy-bundle-docker-0
# For RHOSP16 onwards
podman restart haproxy-bundle-podman-0
Clean Trilio Keystone resources
Trilio registers services and users in Keystone. Those need to be unregistered and deleted.
openstack service delete dmapi
openstack user delete dmapi
Clean Trilio database resources
Trilio creates a database for the dmapi service. This database needs to be cleaned.
Login into the database cluster
## On RHOSP13, run following command on node where database service runs
docker exec -ti -u root galera-bundle-docker-0 mysql -u root
## On RHOSP16
podman exec -it galera-bundle-podman-0 mysql -u root
Run the following SQL statements to clean the database.
## Clean database
DROP DATABASE dmapi;
## Clean dmapi user
=> List 'dmapi' user accounts
MariaDB [mysql]> select user, host from mysql.user where user='dmapi';
+-------+-------------+
| user | host |
+-------+-------------+
| dmapi | 172.25.2.10 |
| dmapi | 172.25.2.8 |
+-------+-------------+
2 rows in set (0.00 sec)
=> Delete those user accounts
MariaDB [mysql]> DROP USER dmapi@172.25.2.10;
Query OK, 0 rows affected (0.82 sec)
MariaDB [mysql]> DROP USER dmapi@172.25.2.8;
Query OK, 0 rows affected (0.05 sec)
=> Verify that dmapi user got cleaned
MariaDB [mysql]> select user, host from mysql.user where user='dmapi';
Empty set (0.00 sec)
Revert overcloud deploy command
Remove the following entries from roles_data.yaml used in the overcloud deploy command.
OS::TripleO::Services::TrilioDatamoverApi
OS::TripleO::Services::TrilioDatamover
Follow these steps to clean the overcloud deploy command from all Trilio entries.
Remove trilio_env.yaml entry
Remove trilio endpoint map file Replace with original map file if existing
Revert back to original RHOSP Horizon container
Run the cleaned overcloud deploy command.
Destroy the Trilio VM Cluster
List all VMs running on the KVM node
virsh list
Destroy the Trilio VMs
virsh destroy <Trilio VM Name or ID>
Undefine the Trilio VMs
virsh undefine <Trilio VM name>
Delete the TrlioVault VM disk from KVM Host storage
Last updated
Was this helpful?