Infrastructure tutorials
Production-grade guides for Linux, servers, security and performance. Copy-paste commands, multi-distro support, written by engineers who run this in production.
Browse by topic
Linux
System administration, shell scripting, package management
Hosting & Servers
Web servers, reverse proxies, SSL, domains
Security
Firewalls, hardening, encryption, access control
Performance
Caching, optimization, profiling, load testing
Databases
MySQL, PostgreSQL, Redis, backups, replication
Networking
DNS, load balancing, VPN, TCP/IP, routing
DevOps
CI/CD, Docker, Kubernetes, automation
Monitoring
Logging, alerting, metrics, observability
Most viewed
Install and configure Deno for web development with systemd and reverse proxy
hostingInstall and configure Caddy web server with automatic HTTPS and reverse proxy
hostingInstall and configure Uvicorn ASGI server with systemd and reverse proxy for FastAPI applications
hostingInstall and configure Ollama for local AI models on Linux servers
devopsConfigure Linux system time synchronization with chrony and NTP hardening
linuxRecently published
Setup Elasticsearch 8 index lifecycle management for automated log retention and storage optimization
devopsConfigure Kafka Streams state stores and RocksDB optimization for high-performance streaming applications
devopsImplement Django continuous deployment with Git hooks and automated testing
devopsConfigure Django Redis caching and session storage for high-performance web applications
performanceSet up Kafka Streams testing framework with TopologyTestDriver for automated stream processing validation
devopsAutomate backup and restore for Ollama models with systemd timers and shell scripts
Set up automated backup and restore procedures for Ollama AI models using systemd timers, shell scripts, and compression. Includes disaster recovery strategies and monitoring integration for production environments.
Deploy multiple Ollama instances with HAProxy load balancing for high availability AI model serving
Set up multiple Ollama instances behind HAProxy load balancer for high-availability AI model serving with health checks, SSL termination, and automatic failover to eliminate single points of failure.
Install and configure Podman for rootless containers on Linux
Learn to install Podman and configure rootless containers as a secure Docker alternative. Includes Docker Compose migration, systemd integration, and troubleshooting common permission issues.
Install and configure Ollama for local AI models on Linux servers
Set up Ollama to run large language models locally on your Linux server. This tutorial covers installation, GPU acceleration, model deployment, API configuration, and performance optimization.
Need help?
Don't want to manage this yourself?
We handle infrastructure for businesses that depend on uptime. From initial setup to ongoing operations.
Talk to an engineer