~/ai-muninn
blog
github
中
~ / blog
/
tag / ollama
❯
grep -r "#ollama" ~/blog
10 matches
date
read
title
2026-04-08
9m
[Benchmark] Rescuing Gemma 4 31B on a 32GB MacBook Pro: From 1.5 to 12.8 tok/s
#gemma-4
#31b
#m1-max
#ollama
2026-04-08
9m
[Benchmark] 4 Machines, 4 Models, 1 Answer: Memory Decides Everything
#gemma-4
#rtx-5090
#dgx-spark
#gb10
2026-04-07
8m
[Benchmark] Gemma 4 E2B vs E4B: 81 tok/s vs 52 on Three Machines — Bandwidth Is Everything
#gemma-4
#e2b
#e4b
#ollama
2026-04-07
8m
[LLM 101] Ollama vs vLLM: Two Ways to Run AI on Your Own Computer
#ollama
#vllm
#llm
#local-ai
2026-04-05
7m
[Benchmark] vLLM vs Ollama on the Same Model: Why 30% Faster on GB10
#vllm
#ollama
#benchmark
#dgx-spark
2026-03-24
7m
[AI Agent] NemoClaw Without the Cloud: Swapping Nemotron for a Local Ollama Model
#nemoclaw
#openclaw
#openshell
#ollama
2026-03-07
7m
[vLLM] Ollama's KEEP_ALIVE Is Silently Eating Your vLLM Headroom
#vllm
#ollama
#gpu-memory
#dgx-spark
2026-03-05
12m
[vLLM] Qwen3.5-35B at 47 tok/s on a Desktop: Migrating from Ollama to vLLM
#dgx-spark
#gb10
#vllm
#ollama
2026-03-05
10m
[AI Agent] Zero API Cost: Running OpenClaw on DGX Spark + Mac Mini
#openclaw
#ai-agent
#dgx-spark
#mac-mini
2026-02-19
11m
[Benchmark] 8 Models on DGX Spark: Finding the Best Stack for AI Agents
#dgx-spark
#gb10
#ollama
#benchmark
← back to all posts