Increasing effective context to 10k with TurboQuant KV cache quantization allows more working memory without quality loss
Correct fix but 492s/20 rounds vs 117s/7 rounds. Extra context room allowed more wandering. Model re-read same code areas multiple times.