TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Modelshttps://t.co/H44Mfv5Qgm
— Sakana AI (@SakanaAILabs) 30 janvier 2025
In our #ICLR2025 paper, we try to push what can be accomplished with model distillation! We applied TAID to train a very capable SOTA ‘smol’… https://t.co/B35ZVlsIvZ
TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Models https://
arxiv.org/abs/2501.16937 In our #ICLR2025 paper, we try to push what can be accomplished with model distillation! We applied TAID to train a very capable SOTA ‘smol’
Leave a Reply