~ /home/coolthor
ai-muninn
Research notes on AI infrastructure, LLM serving, and autonomous agents. Things that took too long to figure out, written down so you don't have to.
❯ whoami
hardware enthusiast running 120B models at home on DGX Spark
building options trading infrastructure with AI agents
occasionally ships iOS apps
❯ ❯ ls -lt ~/blog | head -5
- 2026-04-08[Benchmark] Rescuing Gemma 4 31B on a 32GB MacBook Pro: From 1.5 to 12.8 tok/s
- 2026-04-08[Benchmark] 4 Machines, 4 Models, 1 Answer: Memory Decides Everything
- 2026-04-08[LLM 101] Dense, MoE, PLE, SSM — Four AI Model Architectures Explained Simply
- 2026-04-07[Benchmark] Gemma 4 E2B vs E4B: 81 tok/s vs 52 on Three Machines — Bandwidth Is Everything
- 2026-04-07[Benchmark] From 19 to 50 tok/s: We Quantized Gemma 4 E4B to NVFP4 Before Anyone Else