Sophilex's Blog
  • Home
  • Archive
  • Category
  • Tags
  • About me
  • Friends
BOND: Aligning LLMs with Best-of-N distillation

BOND: Aligning LLMs with Best-of-N distillation

对Best-of-N的生成结果显式表示成一种策略,并蒸馏给模型,将N次推理成功压缩到一次
2025-08-18
学习笔记
#LLM #KD #RLHF
Evaluating Position Bias in Large Language Model Recommendations

Evaluating Position Bias in Large Language Model Recommendations

推荐任务中,item的输入顺序可能会影响模型推荐结果
2025-08-11
学习笔记
#LLM
DATASET DISTILLATION VIA KNOWLEDGE DISTILLATION: TOWARDS EFFICIENT SELF-SUPERVISED PRETRAINING OF DEEP NETWORKS

DATASET DISTILLATION VIA KNOWLEDGE DISTILLATION: TOWARDS EFFICIENT SELF-SUPERVISED PRETRAINING OF DEEP NETWORKS

利用KD在监督学习与自监督学习之间搭了一座桥,非常巧妙!
2025-08-11
学习笔记
#LLM #KD #Pruning
Distilling the Knowledge in Data Pruning

Distilling the Knowledge in Data Pruning

Dataset Pruning与KD的结合
2025-08-10
学习笔记
#LLM #KD #Compression
DA-KD: Difficulty-Aware Knowledge Distillation for  Efficient Large Language Models

DA-KD: Difficulty-Aware Knowledge Distillation for Efficient Large Language Models

训练时,根据模型学习效果动态筛选数据集,dataset efficient
2025-08-04
学习笔记
#LLM #KD
Boosting Parameter Efficiency in LLM-Based Recommendation through Sophisticated Pruning

Boosting Parameter Efficiency in LLM-Based Recommendation through Sophisticated Pruning

针对Rec任务的LLM Pruning策略
2025-08-04
学习笔记
#LLM #Pruning
C2KD: Cross-layer and Cross-head Knowledge Distillation for  Small Language Model-based Recommendations

C2KD: Cross-layer and Cross-head Knowledge Distillation for Small Language Model-based Recommendations

LLM Rec 关于KD的优化
2025-08-04
学习笔记
#LLM #KD
SVD Decompositon in LLM Compression

SVD Decompositon in LLM Compression

事情似乎并没有那么简单...
2025-07-15
slides
#LLM #Compression #SVD
DipSVD: Dual-importance Protected SVD for Efficient LLM Compression

DipSVD: Dual-importance Protected SVD for Efficient LLM Compression

根据每一层的重要性和可压缩程度自适应地分配矩阵压缩率
2025-07-07
学习笔记
#LLM #Matrix_Decomposition
SVD-LLM: TRUNCATION-AWARE SINGULAR VALUE DECOMPOSITION FOR LARGE LANGUAGE MODEL COMPRESSION

SVD-LLM: TRUNCATION-AWARE SINGULAR VALUE DECOMPOSITION FOR LARGE LANGUAGE MODEL COMPRESSION

引入Cholesky decomposition,从理论上保证丢弃的奇异值与loss值一一对应
2025-07-07
学习笔记
#LLM #KD
123456

搜索

Hexo Fluid
京ICP证123456号 | police-icon 京公网安备12345678号
载入天数... 载入时分秒...