Infrastructure tutorials

Production-grade guides for Linux, servers, security and performance. Copy-paste commands, multi-distro support, written by engineers who run this in production.

devops Intermediate

Configure Kubernetes vertical pod autoscaler for resource optimization and cost management

Set up VPA to automatically adjust CPU and memory requests for your Kubernetes workloads. Reduce resource waste and optimize costs by letting VPA analyze actual usage patterns and rightsizing containers.

25 min 4 distros 5 views
devops Advanced

Implement Kubernetes workload rightsizing with VPA recommendations and cost analysis

Set up Vertical Pod Autoscaler to automatically optimize resource requests and limits for your Kubernetes workloads. Create cost analysis dashboards to track resource utilization and identify opportunities for rightsizing containers in production clusters.

45 min 4 distros 44 views
devops Advanced

Configure Kubernetes cluster autoscaler with mixed instance types for cost optimization

Set up Kubernetes cluster autoscaler 1.30 with mixed instance types and spot instances to automatically scale nodes based on demand while minimizing infrastructure costs through intelligent instance selection and workload optimization.

45 min 4 distros 43 views
devops Advanced

Set up Kubernetes custom metrics autoscaling with Prometheus adapter for application-specific scaling

Configure Prometheus adapter to expose custom application metrics to Kubernetes Horizontal Pod Autoscaler for intelligent scaling based on business metrics like queue depth, response time, and user load instead of basic CPU/memory usage.

45 min 4 distros 59 views
devops Intermediate

Implement Kubernetes cluster autoscaler for automatic node scaling

Configure Kubernetes cluster autoscaler to automatically add and remove worker nodes based on pod resource demands. This tutorial covers cloud provider integration, scaling policies, and monitoring for production-grade horizontal scaling.

45 min 4 distros 133 views
devops Intermediate

Configure Kubernetes horizontal pod autoscaler for dynamic scaling based on resource metrics

Set up HPA with CPU and memory targets for automatic pod scaling. Configure metrics server and Prometheus adapter for custom metrics monitoring. Enable dynamic workload scaling based on resource utilization.

45 min 4 distros 221 views

Need help?

Don't want to manage this yourself?

We handle infrastructure for businesses that depend on uptime. From initial setup to ongoing operations.

Talk to an engineer