UI-TARS: Pioneering Automated GUI Interaction with Native Agents Paper • 2501.12326 • Published 13 days ago • 48
ARWKV: Pretrain is not what we need, an RNN-Attention-Based Language Model Born from Transformer Paper • 2501.15570 • Published 8 days ago • 22
Towards General-Purpose Model-Free Reinforcement Learning Paper • 2501.16142 • Published 7 days ago • 23
Over-Tokenized Transformer: Vocabulary is Generally Worth Scaling Paper • 2501.16975 • Published 7 days ago • 20
Optimizing Large Language Model Training Using FP4 Quantization Paper • 2501.17116 • Published 6 days ago • 29
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training Paper • 2501.17161 • Published 6 days ago • 89
Large Language Models Think Too Fast To Explore Effectively Paper • 2501.18009 • Published 5 days ago • 20
Streaming DiLoCo with overlapping communication: Towards a Distributed Free Lunch Paper • 2501.18512 • Published 4 days ago • 22
Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs Paper • 2501.18585 • Published 4 days ago • 40
Reward-Guided Speculative Decoding for Efficient LLM Reasoning Paper • 2501.19324 • Published 3 days ago • 29
HIGGS Collection Models prequantized with [HIGGS](https://arxiv.org/abs/2411.17525) zero-shot quantization. Requires the latest `transformers` to run. • 17 items • Updated Dec 24, 2024 • 6