Update README.md
Browse files
README.md
CHANGED
@@ -6,7 +6,7 @@ license: cc-by-nc-sa-4.0
|
|
6 |
</p>
|
7 |
|
8 |
# Model Card for Taigi-Llama-2-7B
|
9 |
-
Taigi-Llama-2 series are built based on the Traditional Chinese version of the LLaMA-2 model. We conducted continued pre-training on web-scraped data in Taiwanese Hokkien, including Hanzi, POJ, and Hanlo, totaling around 78MB.
|
10 |
|
11 |
For more details, please refer to our [GitHub repository](https://github.com/lbh0830/TW-Hokkien-LLM/tree/main) and the paper: [Enhancing Taiwanese Hokkien Dual Translation by Exploring and Standardizing of Four Writing Systems](https://arxiv.org/abs/2403.12024)
|
12 |
|
@@ -21,7 +21,7 @@ Explore other models and datasets in the [Taiwanese Hokkien LLM collection](http
|
|
21 |
- **Model Size:** 7B parameters
|
22 |
|
23 |
## Usage Example
|
24 |
-
```
|
25 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextGenerationPipeline
|
26 |
import torch
|
27 |
import accelerate
|
|
|
6 |
</p>
|
7 |
|
8 |
# Model Card for Taigi-Llama-2-7B
|
9 |
+
The Taigi-Llama-2 series are built based on the Traditional Chinese version of the LLaMA-2 model. We conducted continued pre-training on web-scraped data in Taiwanese Hokkien, including Hanzi, POJ, and Hanlo, totaling around 78MB.
|
10 |
|
11 |
For more details, please refer to our [GitHub repository](https://github.com/lbh0830/TW-Hokkien-LLM/tree/main) and the paper: [Enhancing Taiwanese Hokkien Dual Translation by Exploring and Standardizing of Four Writing Systems](https://arxiv.org/abs/2403.12024)
|
12 |
|
|
|
21 |
- **Model Size:** 7B parameters
|
22 |
|
23 |
## Usage Example
|
24 |
+
```python
|
25 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextGenerationPipeline
|
26 |
import torch
|
27 |
import accelerate
|