turbo4 K produces garbage because Q pre-rotation guard only checks TURBO3_0, not TURBO4_0
In fattn.cu, turbo_kv only checked GGML_TYPE_TURBO3_0, not TURBO4_0. This gated Q pre-rotation — turbo4 K stored rotated but Q never got pre-rotated, producing garbage dot products (PPL 33K). Fix changes guard to turbo_k_any = (TURBO3_0 || TURBO4_0). After fix: head_dim=256 turbo4 K+V BEATS q8_0 at -0.32%. head_dim=128 turbo4-V excellent (+1.9%), turbo4-K weaker (+5.1%). turbo4 K+V is the BEST quantization option on head_dim=256 models.