共计 52 篇文章
2025
DATASET DISTILLATION VIA KNOWLEDGE DISTILLATION: TOWARDS EFFICIENT SELF-SUPERVISED PRETRAINING OF DEEP NETWORKS
Distilling the Knowledge in Data Pruning
DA-KD: Difficulty-Aware Knowledge Distillation for Efficient Large Language Models
Boosting Parameter Efficiency in LLM-Based Recommendation through Sophisticated Pruning
C2KD: Cross-layer and Cross-head Knowledge Distillation for Small Language Model-based Recommendations
SVD Decompositon in LLM Compression
DipSVD: Dual-importance Protected SVD for Efficient LLM Compression
SVD-LLM: TRUNCATION-AWARE SINGULAR VALUE DECOMPOSITION FOR LARGE LANGUAGE MODEL COMPRESSION
LANGUAGE MODEL COMPRESSION WITH WEIGHTED LOW-RANK FACTORIZATION
ASVD: ACTIVATION-AWARE SINGULAR VALUE DECOMPOSITION FOR COMPRESSING LARGE LANGUAGE MODELS