Instruction tuning has a hidden cost:
✅ Better at following instructions
❌ Narrower output distribution
❌ Worse in-context steerability
We built 🌈 Spectrum Suite to investigate this
and 🌈 Spectrum Tuning as an alternative post-training method —
🤖➡️📉 Post-training made LLMs better at chat and reasoning—but worse at distributional alignment, diversity, and sometimes even steering(!)
We measure this with our new resource (Spectrum Suite) and introduce Spectrum Tuning (method) to bring them back into our models! 🌈
1/🧵