This repo is just a scratchpad for running the Unslopper model locally with llama.cpp.
- Install llama.cpp via Homebrew.
- Use the GGUF Q8_0 quant (the bf16 repo is too large for typical Mac memory).
- Use
llama-completionwith a Qwen-style chat template prompt file and--no-conversationso it does a single completion and exits. - Redirect logs away from stdout so the output file is clean.