Is Multilingual LLM Watermarking Truly Multilingual? A Simple Back-Translation Solution Paper • 2510.18019 • Published Oct 20 • 17
POWSM: A Phonetic Open Whisper-Style Speech Foundation Model Paper • 2510.24992 • Published Oct 28 • 2
Ming-Flash-Omni: A Sparse, Unified Architecture for Multimodal Perception and Generation Paper • 2510.24821 • Published Oct 28 • 37
Generalization or Memorization: Dynamic Decoding for Mode Steering Paper • 2510.22099 • Published Oct 25 • 3
Omni-Reward: Towards Generalist Omni-Modal Reward Modeling with Free-Form Preferences Paper • 2510.23451 • Published Oct 27 • 26
ARC-Encoder: learning compressed text representations for large language models Paper • 2510.20535 • Published Oct 23 • 5
Can Visual Input Be Compressed? A Visual Token Compression Benchmark for Large Multimodal Models Paper • 2511.02650 • Published Nov 4 • 9
RADLADS: Rapid Attention Distillation to Linear Attention Decoders at Scale Paper • 2505.03005 • Published May 5 • 36