Posts

Rust Community Tools You Should Use

Discover essential Rust community tools: Cargo for package management, rustfmt for code formatting, Clippy for linting, and rust-analyzer for language support. Learn how to boost development efficiency and code quality in Rust projects. Rust Community Tools You Should Use

Best Python Tools for Building AI Content Generators

Discover the best Python tools for building AI content generators, including NLP libraries, deep learning frameworks, optimization tools, and deployment solutions for scalable, ethical AI applications. Best Python Tools for Building AI Content Generators

Remote Ollama access via Tailscale or WireGuard, no public ports

Patterns for running Ollama on a home lab or office box and reaching it safely from remote devices. Covers OLLAMA_HOST binding, Tailscale or WireGuard, firewall pinning, and a tight security checklist. Remote Ollama access via Tailscale or WireGuard, no public ports

Go Project Structure: Practices & Patterns

Master Go project layouts with proven patterns from flat structures to hexagonal architecture. Learn when to use cmd/, internal/, pkg/ and avoid common pitfalls. Go Project Structure: Practices & Patterns

Building REST APIs in Go: Complete Guide

A comprehensive guide to implementing RESTful APIs in Go, covering standard library approaches, frameworks, authentication, testing patterns, and production-ready best practices for scalable backend services. Building REST APIs in Go: Complete Guide

Structured Logging in Go with slog for Observability and Alerting

Structured logs turn Go application output into queryable events. Explore log/slog records, JSON handlers, context and trace correlation, redaction, and log-based signals that support monitoring and alerting. Structured Logging in Go with slog for Observability and Alerting

LLM hosting, performance, RAG, and observability

New u pdates of pillar hubs on glukhov.org: organise LLM hosting, performance, RAG, and observability - with dives on runtimes, benchmarks, retrieval, and inference monitoring. https://glukhov.au/posts/2026/llms-hosting-performance-rag-observability #AI #LLM #RAG #Observability #Performance #SelfHosting

Ollama in Docker Compose with GPU and Persistent Model Storage

Run Ollama as a reproducible single-node LLM server using Docker Compose. Configure OLLAMA_HOST and OLLAMA_MODELS, keep models on persistent volumes, enable NVIDIA GPUs, and upgrade safely with rollbacks. Ollama in Docker Compose with GPU and Persistent Model Storage

Ollama behind a reverse proxy with Caddy or Nginx for HTTPS streaming

Expose Ollama securely behind Caddy or Nginx with automated HTTPS, optional Basic Auth or SSO front gates, and correct streaming and WebSocket proxying. Includes timeouts, buffering pitfalls, rate limits, and curl checks. Ollama behind a reverse proxy with Caddy or Nginx for HTTPS streaming