Scaling V norm by constant alpha after quantization compensates for systematic shrinkage and improves quality
MAJOR SUCCESS. Alpha=1.20 improves PPL 5-14% at ALL context lengths. V scaling contributes 6.5x more than K scaling to PPL improvement — consistent with theory (V errors are linear in output while K errors are exponentially amplified by softmax, but V norm directly scales output magnitude). Quantization causes systematic shrinkage of V norms due to codebook discretization, and alpha corrects this. WARNING: PPL-optimal alpha WORSENS KLD — see EXP-0043 for the full picture. Alpha must be validated with KLD, not just PPL.