Sophilex's Blog
  • Home
  • Archive
  • Category
  • Tags
  • About me
  • Friends
学习笔记 25
Importance-Aware Data Selection for Efficient LLM Instruction Tuning Scaling Up Dataset Distillation to ImageNet-1K with Constant Memory Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective FROM CORRECTION TO MASTERY: REINFORCED DISTILLATION OF LARGE LANGUAGE MODEL AGENTS Merge-of-Thought Distillation Delta Knowledge Distillation for Large Language Models Massive Activations in Large Language Models TD3: Tucker Decomposition Based Dataset Distillation Method for Sequential Recommendation Dataset Condensation for Recommendation BOND: Aligning LLMs with Best-of-N distillation More...
slides 5
Different Designs For LLM KD Loss Training-Inference Mismatch In LLM KD(II) SVD Decompositon in LLM Compression Training-Inference Mismatch In LLM KD Different Designs For LLM KD Loss
Machine Learning 3
SGD收敛性学习笔记 Reparameterization-Trick Attention-Machanism
组合数学 3
Dilworth 定理 多项式EXP运算的组合意义 树与图上计数问题:Prufer序列与LGV引理
抽象代数 3
环论 群作用 Burnside 引理
动态规划 3
高维前缀和学习笔记 各类基于决策单调性的dp优化 斜率优化dp总结
配置笔记 3
hexo+reveal指南 服务器转发流量至本地 服务器炼丹手册
杂想 3
练琴有感 伍六七观后小感 杂想-抽奖问题平凡解
机器学习 2
从REINFORCE到PPO VAE学习笔记
编译原理 2
编译原理实验七-sql简易编译器 编译原理实验六-对表达式的解析
博弈论 1
二分图博弈学习笔记
鞅论 1
势函数与鞅的停时定理
Bugs 1
Bug聚集地

搜索

Hexo Fluid
京ICP证123456号 | police-icon 京公网安备12345678号
载入天数... 载入时分秒...