Slamming: Training a Speech Language Model on One GPU in a Day Paper • 2502.15814 • Published 6 days ago • 37
LLM-Microscope: Uncovering the Hidden Role of Punctuation in Context Memory of Transformers Paper • 2502.15007 • Published 5 days ago • 132
SuperGPQA: Scaling LLM Evaluation across 285 Graduate Disciplines Paper • 2502.14739 • Published 5 days ago • 91
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features Paper • 2502.14786 • Published 5 days ago • 115
view article Article PaliGemma 2 Mix - New Instruction Vision Language Models by Google 7 days ago • 59
Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention Paper • 2502.11089 • Published 9 days ago • 134
ReLearn: Unlearning via Learning for Large Language Models Paper • 2502.11190 • Published 9 days ago • 28
mmE5: Improving Multimodal Multilingual Embeddings via High-quality Synthetic Data Paper • 2502.08468 • Published 13 days ago • 13
InfiniteHiP: Extending Language Model Context Up to 3 Million Tokens on a Single GPU Paper • 2502.08910 • Published 13 days ago • 141
TransMLA: Multi-head Latent Attention Is All You Need Paper • 2502.07864 • Published 14 days ago • 44
Scaling Pre-training to One Hundred Billion Data for Vision Language Models Paper • 2502.07617 • Published 14 days ago • 28