Sophilex's Blog
  • Home
  • Archive
  • Category
  • Tags
  • About me
  • Friends
学习笔记 16
BOND: Aligning LLMs with Best-of-N distillation Evaluating Position Bias in Large Language Model Recommendations DATASET DISTILLATION VIA KNOWLEDGE DISTILLATION: TOWARDS EFFICIENT SELF-SUPERVISED PRETRAINING OF DEEP NETWORKS Distilling the Knowledge in Data Pruning DA-KD: Difficulty-Aware Knowledge Distillation for Efficient Large Language Models Boosting Parameter Efficiency in LLM-Based Recommendation through Sophisticated Pruning C2KD: Cross-layer and Cross-head Knowledge Distillation for Small Language Model-based Recommendations DipSVD: Dual-importance Protected SVD for Efficient LLM Compression SVD-LLM: TRUNCATION-AWARE SINGULAR VALUE DECOMPOSITION FOR LARGE LANGUAGE MODEL COMPRESSION LANGUAGE MODEL COMPRESSION WITH WEIGHTED LOW-RANK FACTORIZATION More...
Machine Learning 3
SGD收敛性学习笔记 Reparameterization-Trick Attention-Machanism
组合数学 3
Dilworth 定理 多项式EXP运算的组合意义 树与图上计数问题:Prufer序列与LGV引理
抽象代数 3
环论 群作用 Burnside 引理
slides 3
SVD Decompositon in LLM Compression Training-Inference Mismatch In LLM KD Different Designs For LLM KD Loss
动态规划 3
高维前缀和学习笔记 各类基于决策单调性的dp优化 斜率优化dp总结
配置笔记 3
hexo+reveal指南 服务器转发流量至本地 服务器炼丹手册
机器学习 2
从REINFORCE到PPO VAE学习笔记
杂想 2
练琴有感 杂想-抽奖问题平凡解
编译原理 2
编译原理实验七-sql简易编译器 编译原理实验六-对表达式的解析
bugs 1
linux 1
bug聚集地
博弈论 1
二分图博弈学习笔记
鞅论 1
势函数与鞅的停时定理

搜索

Hexo Fluid
京ICP证123456号 | police-icon 京公网安备12345678号
载入天数... 载入时分秒...