共计 30 篇文章
2025
Delta Knowledge Distillation for Large Language Models
Massive Activations in Large Language Models
TD3: Tucker Decomposition Based Dataset Distillation Method for Sequential Recommendation
Dataset Condensation for Recommendation
BOND: Aligning LLMs with Best-of-N distillation
Evaluating Position Bias in Large Language Model Recommendations
DATASET DISTILLATION VIA KNOWLEDGE DISTILLATION: TOWARDS EFFICIENT SELF-SUPERVISED PRETRAINING OF DEEP NETWORKS
Distilling the Knowledge in Data Pruning
DA-KD: Difficulty-Aware Knowledge Distillation for Efficient Large Language Models
Boosting Parameter Efficiency in LLM-Based Recommendation through Sophisticated Pruning