Sophilex's Blog
  • Home
  • Archive
  • Category
  • Tags
  • About me
  • Friends

共计 60 篇文章


2025

12-29
Distilling the Essence: Efficient Reasoning Distillation via Sequence Truncation
12-29
Instruction tuning with loss over instructions
12-15
Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning
12-14
RETAINING BY DOING: THE ROLE OF ON-POLICY DATA IN MITIGATING FORGETTING
12-08
Proximal Gradient and Subgradients
11-17
Different Designs For LLM KD Loss(II)
11-17
Importance-Aware Data Selection for Efficient LLM Instruction Tuning
10-13
Scaling Up Dataset Distillation to ImageNet-1K with Constant Memory
10-13
Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective
10-11
Training-Inference Mismatch In LLM KD(II)
123…6

搜索

Hexo Fluid
京ICP证123456号 | police-icon 京公网安备12345678号
载入天数... 载入时分秒...