Infrastructure tutorials
Production-grade guides for Linux, servers, security and performance. Copy-paste commands, multi-distro support, written by engineers who run this in production.
Browse by topic
Linux
System administration, shell scripting, package management
Hosting & Servers
Web servers, reverse proxies, SSL, domains
Security
Firewalls, hardening, encryption, access control
Performance
Caching, optimization, profiling, load testing
Databases
MySQL, PostgreSQL, Redis, backups, replication
Networking
DNS, load balancing, VPN, TCP/IP, routing
DevOps
CI/CD, Docker, Kubernetes, automation
Monitoring
Logging, alerting, metrics, observability
Most viewed
Install and configure ArgoCD for GitOps continuous deployment with RBAC and SSL
devopsInstall and configure CockroachDB cluster with high availability and distributed SQL
databasesInstall and configure PostgreSQL 17 with performance tuning and security hardening
databasesInstall and configure Loki for centralized log aggregation with Grafana integration
monitoringInstall and configure WireGuard VPN server with client management
networkingRecently published
Configure Nagios SNMP monitoring for network devices with automated discovery and templates
monitoringIntegrate Elasticsearch 8 with Prometheus monitoring and Grafana dashboards
monitoringSet up NGINX web application firewall with ModSecurity 3 and OWASP Core Rule Set
securityConfigure NGINX reverse proxy with advanced caching and load balancing
hostingConfigure Netdata alerts with Slack and Microsoft Teams for real-time monitoring notifications
monitoringImplement Apache Spark 3.5 cluster with YARN and HDFS for distributed computing
Set up a production-grade Apache Spark 3.5 cluster with YARN resource management and HDFS distributed storage for scalable big data processing. This tutorial covers multi-node Hadoop cluster configuration, YARN integration, and monitoring setup.
Set up Spark 3.5 Delta Lake with MinIO for ACID transactions and big data analytics
Configure Apache Spark 3.5 with Delta Lake and MinIO object storage for ACID transactions, data versioning, and distributed analytics processing. Includes complete setup for production-grade data lake architecture.
Implement Spark streaming with Kafka and MinIO for real-time analytics and big data processing
Build a production-ready real-time analytics pipeline using Apache Spark 3.5 streaming, Kafka for data ingestion, and MinIO for distributed object storage. This tutorial covers fault-tolerant streaming configurations and end-to-end pipeline implementation.
Configure MinIO with Apache Spark 3.5 for big data analytics and object storage integration
Set up Apache Spark 3.5 with MinIO S3-compatible object storage for scalable big data analytics. Configure distributed storage, implement data lake patterns, and run production analytics workflows on your cluster infrastructure.
Need help?
Don't want to manage this yourself?
We handle infrastructure for businesses that depend on uptime. From initial setup to ongoing operations.
Talk to an engineer