Posts

How to Self-Host Ollama and Open WebUI with NVIDIA GPU on Ubuntu 22.04/24.04

Run a Local LLM with GPU Acceleration: Deploy Ollama + Open WebUI on Ubuntu via Docker

Run MinIO with Docker and Caddy: Secure S3-Compatible Object Storage on Ubuntu 24.04