NanoBEIR 🍺 Collection A collection of smaller versions of BEIR datasets with 50 queries and up to 10K documents each. • 13 items • Updated Sep 11, 2024 • 12
view article Article KV Caching Explained: Optimizing Transformer Inference Efficiency By not-lain • 5 days ago • 22
Towards General Text Embeddings with Multi-stage Contrastive Learning Paper • 2308.03281 • Published Aug 7, 2023 • 2
Jasper and Stella: distillation of SOTA embedding models Paper • 2412.19048 • Published Dec 26, 2024 • 1
Facilitating large language model Russian adaptation with Learned Embedding Propagation Paper • 2412.21140 • Published Dec 30, 2024 • 16
Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks Paper • 1908.10084 • Published Aug 27, 2019 • 5
Smarter, Better, Faster, Longer: A Modern Bidirectional Encoder for Fast, Memory Efficient, and Long Context Finetuning and Inference Paper • 2412.13663 • Published Dec 18, 2024 • 125
Tulu 3 Models Collection All models released with Tulu 3 -- state of the art open post-training recipes. • 10 items • Updated 5 days ago • 84
On the Power of Decision Trees in Auto-Regressive Language Modeling Paper • 2409.19150 • Published Sep 27, 2024 • 4
AutoTrain: No-code training for state-of-the-art models Paper • 2410.15735 • Published Oct 21, 2024 • 59
PingPong: A Benchmark for Role-Playing Language Models with User Emulation and Multi-Model Evaluation Paper • 2409.06820 • Published Sep 10, 2024 • 64
Building and better understanding vision-language models: insights and future directions Paper • 2408.12637 • Published Aug 22, 2024 • 124
LLM Pruning and Distillation in Practice: The Minitron Approach Paper • 2408.11796 • Published Aug 21, 2024 • 58
To Code, or Not To Code? Exploring Impact of Code in Pre-training Paper • 2408.10914 • Published Aug 20, 2024 • 42
The Russian-focused embedders' exploration: ruMTEB benchmark and Russian embedding model design Paper • 2408.12503 • Published Aug 22, 2024 • 24