This sample demonstrates how to use KEDA to scale a Kubernetes deployment based on multiple metrics. Specifically for not scaling up deployment when the error threshold is exceeded, else scaling based on a queue length metric. This keda external scaler implementation scales down 1 instance at a time during scale down.
This keda custom scaler customizations has been written in go. As required with keda external scaler we need to implement a few operations. The key ones being GetMetricSpec , and GetMetrics which returns the customized metric value to keda, based on which the scale up or scale down happens.
The inital installation steps from the prometheus scaler can be reused, other than the last line (Which creates keda scaled object with prometheus trigger). Refer the setup.sh, this script installs the base required components for this sample. Details of the components have been added as comments in the script.
Key points to note:
- After the Line in the setup.sh script, which builds the test go application container image, and pushes it to ACR, you should update the ACR_NAME with your created ACR name in deployment.yaml file, for both the deployments. This deployment will be used in the tests to modify the metrics in prometheus
- if you want to install the external scaler, do not execute the final line of the setup.sh file (Which creates keda scaled object with prometheus trigger).
- The Prometheus scrape config scrapes the test go apps for the msg_queue_length and rate_429_errors metrics.
The additional setup steps for external scaler are in the file additional-setup-for-external-scaler.sh. The $ACR_NAME in the ext-scaler-deployment-service.yaml and to-be-scaled-workload-deployment.yaml files need to be modified with you ACR name.
Details of how the keda scaled object integrates with the external scaler please look at keda-scaled-object-with-ext-scaler.yaml
After the installation, you can follow the steps in tests.sh to test the scaling behavior.