Yiran0924 commited on
Commit
c484065
·
verified ·
1 Parent(s): 554afc9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -51,7 +51,7 @@ tags:
51
  We introduce **Babel**, a multilingual LLM that covers the top 25 languages by number of speakers, including English, Chinese, Hindi, Spanish, Arabic, French, Bengali, Portuguese, Russian, Urdu, Indonesian, German, Japanese, Swahili, Filipino, Tamil, Vietnamese, Turkish, Italian, Javanese, Korean, Hausa, Persian, Thai, and Burmese. These 25 languages support over 90% of the global population, and include many languages neglected by other open multilingual LLMs. Unlike traditional continued pretraining approaches, Babel expands its parameter count through a layer extension technique that elevates Babel's performance ceiling.
52
 
53
  We introduce two variants:
54
- - **Babel-9B**, designed for efficient single-GPU inference and fine-tuning
55
  - **Babel-83B**, which sets a new standard for open multilingual LLMs
56
 
57
  Extensive evaluations on multilingual tasks demonstrate its superior performance compared to open LLMs of comparable size. In addition, using existing supervised fine-tuning datasets, Babel achieves remarkable performance, with **Babel-9B-Chat** leading among 10B-sized LLMs and **Babel-83B-Chat** setting a new standard for open LLMs, performing comparably to GPT-4o on certain tasks.
 
51
  We introduce **Babel**, a multilingual LLM that covers the top 25 languages by number of speakers, including English, Chinese, Hindi, Spanish, Arabic, French, Bengali, Portuguese, Russian, Urdu, Indonesian, German, Japanese, Swahili, Filipino, Tamil, Vietnamese, Turkish, Italian, Javanese, Korean, Hausa, Persian, Thai, and Burmese. These 25 languages support over 90% of the global population, and include many languages neglected by other open multilingual LLMs. Unlike traditional continued pretraining approaches, Babel expands its parameter count through a layer extension technique that elevates Babel's performance ceiling.
52
 
53
  We introduce two variants:
54
+ - **Babel-9B**, designed for efficient inference and fine-tuning
55
  - **Babel-83B**, which sets a new standard for open multilingual LLMs
56
 
57
  Extensive evaluations on multilingual tasks demonstrate its superior performance compared to open LLMs of comparable size. In addition, using existing supervised fine-tuning datasets, Babel achieves remarkable performance, with **Babel-9B-Chat** leading among 10B-sized LLMs and **Babel-83B-Chat** setting a new standard for open LLMs, performing comparably to GPT-4o on certain tasks.