Overview
The following tutorials explain how you can benefit from using Kedify in various use cases:
HTTP Scaling Guides
Section titled “HTTP Scaling Guides”The guides in this section cover various HTTP scaling strategies and configurations, helping you choose the best approach for your application’s deployment scenario.
Basic Scaling Guides
Section titled “Basic Scaling Guides”- HTTP Scaling for In-Cluster Traffic (Kubernetes Service) - For applications that do not expose an external Ingress and rely solely on internal Kubernetes Services.
- HTTP Scaling for Ingress-Based Applications - For applications exposed using Kubernetes Ingress.
- HTTP Scaling for Ingress-Based Inference Workloads - For inference workloads exposed using Kubernetes Ingress.
- HTTP Scaling for OpenShift-Based Applications - For applications exposed using OpenShift Routes.
- HTTP Scaling with Kubernetes Gateway API - For applications exposed via the Gateway API, Kedify utilizes its autowiring featurel.
- HTTP Scaling with Istio VirtualServices - This guide demonstrates how to scale applications within an Istio service mesh using
VirtualServiceresources to manage HTTP traffic. - HTTP Scaling with TLS for Ingress-Based Applications - For TLS workloads using Ingress, Kedify intercepts traffic and collects encrypted traffic metrics without interfering with your TLS setup.
Advanced Configuration
Section titled “Advanced Configuration”- Configure Waiting and Maintenance Pages for HTTP Scaler - Configuration options to set static pages when your application is going through cold-starts or needs temporarily re-route the traffic.
- Configure Envoy in the Kedify Proxy - How to configure Kedify Proxy Envoy for advanced use cases with custom Envoy config snippets.
- Kedify Proxy Performance Tuning - For applications that use Kedify Proxy to autoscale based on HTTP traffic.
- OpenTelemetry Tracing for Kedify Proxy & Interceptor - Enable OpenTelemetry tracing to get insights about HTTP calls handled by Kedify Proxy and HTTP Add-on interceptor components.
Custom Metrics and OTel Integration
Section titled “Custom Metrics and OTel Integration”- Scaling with Custom Metrics and OTel Collectors - Suitable for applications that expose custom metrics either as a scrapeable metrics endpoint or as a OTLP exporter. Example application is AI workload (namely vLLM with LLama 3.1)
- Custom Metrics with Prometheus Scaler & Migration to OTel Scaler - Setting up simple demo app that exposes application-specific metrics, Prometheus Scaler and then migrating everything to OTel Scaler.
- Kedify OTel Scaler Integrations - Patterns - Multiple ways to integrate your workloads with Kedify OTel Scaler.
Predictive Scaling
Section titled “Predictive Scaling”- Enabling Predictive Scaling in Kedify Installation - Describing ways to deploy Kedify Predictor using the web UI as well as the helm installation.
- Predictive Scaling Example - A tutorial how to set up a predictive scaler that internally trains a model that learns from historical data and suggests unseen values for e-commerce data (number of incoming request in e-shop).
- Predictor & OTLP Example - Prometheus exporter scrapes periodically the weather data, OTel collector forwards this to Kedify Predictor that learns from the data and suggests the future values.
Vertical Scaling
Section titled “Vertical Scaling”- PodResourceProfile Reacting on a ScaledObject - Shrink the last replica to lower resources when the
ScaledObjectis passive. - Vertical Scaling with PodResourceAutoscaler - Continuously scale CPU and memory requests/limits in place based on changing runtime load.