Homelab

list by @rakhim

How I Self-Hosted Llama 3.2 with Coolify on My Home Server: A Step-by-Step Guide
GEEK.SG | 16 Oct 2024 | original ↗

Inspired by numerous people migrating their Next.js applications from Vercel to self-hosted VPS on Hetzner due to pricing concerns, I decided to explore self-hosting some of my non-critical applications. Additionally, I wanted to push my technical boundaries by running Llama 3.2 using Ollama and making its API available to power AI applications...