Sophilex's Blog
  • Home
  • Archive
  • Category
  • Tags
  • About me
  • Friends
Importance-Aware Data Selection for Efficient LLM Instruction Tuning

Importance-Aware Data Selection for Efficient LLM Instruction Tuning

llm微调时的一种数据选择策略
2025-11-17
学习笔记
#LLM
Scaling Up Dataset Distillation to ImageNet-1K with Constant Memory

Scaling Up Dataset Distillation to ImageNet-1K with Constant Memory

优化MTT,使其可以scale到更大规模的cv数据集上
2025-10-13
学习笔记
#KD #Dataset_Condensation
Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective

Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective

对传统数据集蒸馏的双层优化结构进行解耦,实现线性复杂度
2025-10-13
学习笔记
#KD #Dataset_Condensation
Training-Inference Mismatch In LLM KD(II)

Training-Inference Mismatch In LLM KD(II)

组会汇报
2025-10-11
slides
#LLM #KD
FROM CORRECTION TO MASTERY: REINFORCED DISTILLATION OF LARGE LANGUAGE MODEL AGENTS

FROM CORRECTION TO MASTERY: REINFORCED DISTILLATION OF LARGE LANGUAGE MODEL AGENTS

学生生成SGO时,教师在必要时给予干预,压缩理论误差上界
2025-09-28
学习笔记
#LLM #KD
Merge-of-Thought Distillation

Merge-of-Thought Distillation

KD时,对不同教师来源的信息做了一个巧妙的融合处理
2025-09-28
学习笔记
#LLM #KD
Delta Knowledge Distillation for Large Language Models

Delta Knowledge Distillation for Large Language Models

将KD的对齐目标变为模型更新的变化量,而不是固定的token概率分布
2025-09-28
学习笔记
#LLM #KD
Massive Activations in Large Language Models

Massive Activations in Large Language Models

关于LLM中的异常大激活值,以及其与attention sink的关系,将其解释为模型中统一且固定的bias,挺有意思的
2025-09-21
学习笔记
#LLM
TD3: Tucker Decomposition Based Dataset Distillation Method for Sequential Recommendation

TD3: Tucker Decomposition Based Dataset Distillation Method for Sequential Recommendation

序列推荐数据集的蒸馏,引入Tucker分解来缓解随数据集规模而增长的参数压力
2025-09-21
学习笔记
#Dataset_Condensation
Dataset Condensation for Recommendation

Dataset Condensation for Recommendation

推荐数据集蒸馏
2025-09-14
学习笔记
#KD #Dataset_Condensation
1234…6

搜索

Hexo Fluid
京ICP证123456号 | police-icon 京公网安备12345678号
载入天数... 载入时分秒...