Sophilex's Blog
  • Home
  • Archive
  • Category
  • Tags
  • About me
  • Friends
学习笔记 24
Scaling Up Dataset Distillation to ImageNet-1K with Constant Memory Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective FROM CORRECTION TO MASTERY: REINFORCED DISTILLATION OF LARGE LANGUAGE MODEL AGENTS Merge-of-Thought Distillation Delta Knowledge Distillation for Large Language Models Massive Activations in Large Language Models TD3: Tucker Decomposition Based Dataset Distillation Method for Sequential Recommendation Dataset Condensation for Recommendation BOND: Aligning LLMs with Best-of-N distillation Evaluating Position Bias in Large Language Model Recommendations More...
Machine Learning 3
SGD收敛性学习笔记 Reparameterization-Trick Attention-Machanism
组合数学 3
Dilworth 定理 多项式EXP运算的组合意义 树与图上计数问题:Prufer序列与LGV引理
抽象代数 3
环论 群作用 Burnside 引理
slides 3
SVD Decompositon in LLM Compression Training-Inference Mismatch In LLM KD Different Designs For LLM KD Loss
动态规划 3
高维前缀和学习笔记 各类基于决策单调性的dp优化 斜率优化dp总结
配置笔记 3
hexo+reveal指南 服务器转发流量至本地 服务器炼丹手册
机器学习 2
从REINFORCE到PPO VAE学习笔记
杂想 2
练琴有感 杂想-抽奖问题平凡解
编译原理 2
编译原理实验七-sql简易编译器 编译原理实验六-对表达式的解析
bugs 1
linux 1
bug聚集地
博弈论 1
二分图博弈学习笔记
鞅论 1
势函数与鞅的停时定理

搜索

Hexo Fluid
京ICP证123456号 | police-icon 京公网安备12345678号
载入天数... 载入时分秒...