ResQ: Mixed-Precision Quantization of Large Language Models with Low-Rank Residuals Paper • 2412.14363 • Published Dec 18, 2024 • 1
TRIM: Token-wise Attention-Derived Saliency for Data-Efficient Instruction Tuning Paper • 2510.07118 • Published Oct 8, 2025 • 10