Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
ehristoforu 
posted an update 1 day ago
Post
1886
Introducing our first standalone model – FluentlyLM Prinum

Introducing the first standalone model from Project Fluently LM! We worked on it for several months, used different approaches and eventually found the optimal one.

General characteristics:
- Model type: Causal language models (QwenForCausalLM, LM Transformer)
- Number of parameters: 32.5B
- Number of parameters (not embedded): 31.0B
- Number of layers: 64
- Context: 131,072 tokens
- Language(s) (NLP): English, French, Spanish, Russian, Chinese, Japanese, Persian (officially supported)
- License: MIT

Creation strategy:
The basis of the strategy is shown in Pic. 2.
We used Axolotl & Unsloth for SFT-finetuning with PEFT LoRA (rank=64, alpha=64) and Mergekit for SLERP and TIES mergers.

Evolution:
🏆 12th place in the Open LLM Leaderboard ( open-llm-leaderboard/open_llm_leaderboard) (21.02.2025)

Detailed results and comparisons are presented in Pic. 3.

Links:
- Model: fluently-lm/FluentlyLM-Prinum
- GGUF version: mradermacher/FluentlyLM-Prinum-GGUF
- Demo on ZeroGPU: ehristoforu/FluentlyLM-Prinum-demo

Sounds great! Thanks for making it free software, free as in freedom, that creators can use it as they wish and want, modify, and build on it.

What is Free Software? - GNU Project - Free Software Foundation
https://www.gnu.org/philosophy/free-sw.html

Let me try it out. I hope it runs on my ROG Strix GeForce RTX 3090 24 GB

·

Thank you for your interest in our model! I hope you like it)