T4K Resource Usage

This page provides a guideline on how to calculate the resource needs under various circumstances

Deprecated Documentation

This document is deprecated and no longer supported. For accurate, up-to-date information, please refer to the documentation for the latest version of Trilio.

# T4K Resource Usage

In order to provide hardware recommendations to users, a series of tests were performed on T4K to measure the memory usage of the control plane, analyzer, and web-backend resources. The tests were performed mainly based on:

  • Number of Kubernetes resources

  • Number of backup plans

Resource Based Consumption

In order to analyze the impact of resources on T4K, a T4K was set up with 1000 active namespace level backups, 50 backup plans (not running in schedule or creating backups), and 10k resources which contained 1 deployment, 10k services, 1500 config maps, and 750 secrets initially. In each iteration of tests, the backups were kept at 1000, and 10k resources were added until the cluster had 50k resources and 1000 backups.

Below table and graphs provide memory insights into the control plane, analyzer, and web backend:

Metrics

Load
Control Plane
Analyzer
Web Backend

Resources/Backups

Memory (idle)

Memory (spike)

Memory (idle)

Memory (spike)

Memory (idle)

Memory (spike)

10k resources (1000 backups)

263

263

107

107

272

443

20k resources (1000 backups)

320

320

102

102

344

575

30k resources (1000 backups)

342

480

104

107

398

1130

40k resources (1000 backups)

391

553

111

159

443

1250

50k resources (1000 backups)

462

650

140

179

515

1640

50k resources (1000 backups) - after restart

449

671

142

142

483

1370

Chart

Resource Usage of Control Plane
Resource Usage of Analyzer
Resource Usage of Web Backend

Adding 10k resources consumes memory usage up to 150 MB, 52 MB, and 555 MB in the control plane, analyzer, and web-backend respectively.

Resource Consumption During Backups

In order to analyze the impact of backups on T4K, initially, a T4K was setup with 10k resources which contained 1 deployment, 10k services, 1500 config maps and 750 secrets, ~50 active backups and 50 backup plans creating backups at the same time. In each iteration of tests, the resources were kept at 10k and 1000 backups were added until the cluster had 6k backups and 10k resources.

Below table and graphs provide memory insights of control-plane, analyzer and web-backend:

Metrics

Load
Control Plane
Analyzer
Web Backend

Resources/Backups

Memory (idle)

Memory (spike)

Memory (idle)

Memory (spike)

Memory (idle)

Memory (spike)

10k resources

(0-50 backups)

330

387

44

51

380

620

10k resources

(1000 backups)

476

626

164

191

457

860

10k resources

(2000 backups)

630

780

210

255

757

1190

10k resources

(3300 backups)

750

960

310

400

912

1490

10k resources

(6000 backups)

1060

1320

796

1002

1030

1700

10k resources

(6000 backups) - after restart

552

-

382

-

558

1300

Chart

Control Plane Resource Consumption with Backups

Note: No spike was seen after the restart in above graph as all scheduled based backups were paused at that time

Analyzer Resource Consumption with Backups

Note: No spike was seen after the restart in above graph as all scheduled based backups were paused at that time

Web Backend Resource Consumption with Backups

Addition of 1k backups consumes memory usage upto 239 MB, 200 MB and 330 MB in control-plane, analyzer and web-backend respectively.

Conclusion

control-plane and analyzer components consume more memory with increasing backups than increasing resources. The number of resources has small amount of effect on control-plane and analyzer components therefore, the number of backups factor should be taken into the account more while providing the memory limits for these containers. The control-plane spikes seem to be related to scheduled backup plans creating 50 backups every hour. The number of backups that are being created at the same time also contributes in memory spike seen in control-plane. The same spike isn't seen much in the resource based tests as the backup plans were not creating backups every hour.

web-backend on the other hand consume memory with both increasing backups and resources as it needs to cache every resources in the cluster.

Considering the metrics from the tests above, below recommendations can be followed to configure control-plane, analyzer and web-backend's memory limits:

Component
Recommendation

control-plane

1 GB for 10k resources, 1k backups

analyzer

512 MB for 10k resources, 1k backups

web-backend

1 GB for 10k resources, 1k backups

Last updated