Skipping V dequant for attention weights below 1e-6 has zero quality impact and improves decode speed
PPL numerically identical ON vs OFF. NIAH improved (9/9 vs 7/9 q8_0 baseline) due to reduced quantization noise on irrelevant positions. Skip rate scales with context length.