Infrastructure tutorials
Production-grade guides for Linux, servers, security and performance. Copy-paste commands, multi-distro support, written by engineers who run this in production.
Browse by topic
Linux
System administration, shell scripting, package management
Hosting & Servers
Web servers, reverse proxies, SSL, domains
Security
Firewalls, hardening, encryption, access control
Performance
Caching, optimization, profiling, load testing
Databases
MySQL, PostgreSQL, Redis, backups, replication
Networking
DNS, load balancing, VPN, TCP/IP, routing
DevOps
CI/CD, Docker, Kubernetes, automation
Monitoring
Logging, alerting, metrics, observability
Most viewed
Install and configure Deno for web development with systemd and reverse proxy
hostingInstall and configure Caddy web server with automatic HTTPS and reverse proxy
hostingInstall and configure Uvicorn ASGI server with systemd and reverse proxy for FastAPI applications
hostingInstall and configure Ollama for local AI models on Linux servers
devopsConfigure Linux system time synchronization with chrony and NTP hardening
linuxRecently published
Setup Elasticsearch 8 index lifecycle management for automated log retention and storage optimization
devopsConfigure Kafka Streams state stores and RocksDB optimization for high-performance streaming applications
devopsImplement Django continuous deployment with Git hooks and automated testing
devopsConfigure Django Redis caching and session storage for high-performance web applications
performanceSet up Kafka Streams testing framework with TopologyTestDriver for automated stream processing validation
devopsConfigure Kubernetes cluster autoscaler with mixed instance types for cost optimization
Set up Kubernetes cluster autoscaler 1.30 with mixed instance types and spot instances to automatically scale nodes based on demand while minimizing infrastructure costs through intelligent instance selection and workload optimization.
Set up Docker Compose monitoring stack with Prometheus and Grafana for AI model performance tracking
Deploy a complete monitoring stack using Docker Compose with Prometheus for metrics collection and Grafana for visualization, specifically configured to track AI model performance metrics like inference latency, throughput, and resource utilization.
Advanced Nomad job templates and deployment strategies with rolling updates and canary deployments
Master production-grade Nomad job templates with HCL syntax, implement rolling updates with health checks, and deploy advanced blue-green and canary deployment patterns for resilient containerized workloads.
Deploy multiple Ollama instances with HAProxy load balancing for high availability AI model serving
Set up multiple Ollama instances behind HAProxy load balancer for high-availability AI model serving with health checks, SSL termination, and automatic failover to eliminate single points of failure.
Install and configure Ollama for local AI models on Linux servers
Set up Ollama to run large language models locally on your Linux server. This tutorial covers installation, GPU acceleration, model deployment, API configuration, and performance optimization.
Need help?
Don't want to manage this yourself?
We handle infrastructure for businesses that depend on uptime. From initial setup to ongoing operations.
Talk to an engineer