Scaling Up Dataset Distillation to ImageNet-1K with Constant Memory 优化MTT,使其可以scale到更大规模的cv数据集上 2025-10-13 学习笔记 #KD #Dataset_Condensation
Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective 对传统数据集蒸馏的双层优化结构进行解耦,实现线性复杂度 2025-10-13 学习笔记 #KD #Dataset_Condensation
FROM CORRECTION TO MASTERY: REINFORCED DISTILLATION OF LARGE LANGUAGE MODEL AGENTS 学生生成SGO时,教师在必要时给予干预,压缩理论误差上界 2025-09-28 学习笔记 #LLM #KD
Delta Knowledge Distillation for Large Language Models 将KD的对齐目标变为模型更新的变化量,而不是固定的token概率分布 2025-09-28 学习笔记 #LLM #KD