dahara1 commited on
Commit
a55b368
1 Parent(s): f62bfb6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -18,6 +18,10 @@ And translation ability for languages other than Japanese and English has deteri
18
  If you want to translate the entire file at once, try Colab below.
19
  [ALMA_7B_Ja_GPTQ_Ja_En_batch_translation_sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_GPTQ_Ja_En_batch_translation_sample.ipynb)
20
 
 
 
 
 
21
 
22
  **ALMA** (**A**dvanced **L**anguage **M**odel-based tr**A**nslator) is an LLM-based translation model, which adopts a new translation model paradigm: it begins with fine-tuning on monolingual data and is further optimized using high-quality parallel data. This two-step fine-tuning process ensures strong translation performance.
23
  Please find more details in their [paper](https://arxiv.org/abs/2309.11674).
 
18
  If you want to translate the entire file at once, try Colab below.
19
  [ALMA_7B_Ja_GPTQ_Ja_En_batch_translation_sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_GPTQ_Ja_En_batch_translation_sample.ipynb)
20
 
21
+ if you enconter error below.
22
+ ```RuntimeError: probability tensor contains either `inf`, `nan` or element < 0```
23
+ It's mean your memory is not enough. decrease your num_beams or token size.
24
+
25
 
26
  **ALMA** (**A**dvanced **L**anguage **M**odel-based tr**A**nslator) is an LLM-based translation model, which adopts a new translation model paradigm: it begins with fine-tuning on monolingual data and is further optimized using high-quality parallel data. This two-step fine-tuning process ensures strong translation performance.
27
  Please find more details in their [paper](https://arxiv.org/abs/2309.11674).