Update README.md
Browse files
README.md
CHANGED
@@ -16,10 +16,13 @@ tags:
|
|
16 |
|
17 |
<strong><span style="font-size: larger;">bertin-gpt-j-6B-alpaca-4bit-128g 🤗</span></strong>
|
18 |
|
|
|
|
|
19 |
**descripción en español agregado ⬇️**
|
20 |
|
21 |
This is a 4-bit GPTQ version of the [bertin-project/bertin-gpt-j-6B-alpaca]( https://huggingface.co/bertin-project/bertin-gpt-j-6B-alpaca)
|
22 |
|
|
|
23 |
this is the result of quantizing to 4 bits using [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ).
|
24 |
|
25 |
** How to easily download and use this model in text-generation-webui** (tutorial by [TheBloke](https://huggingface.co/TheBloke))
|
@@ -61,7 +64,7 @@ This dataset cannot be used to create models that compete in any way with OpenAI
|
|
61 |
Finetuning
|
62 |
To fine-tune the BERTIN GPT-J-6B model we used the code available on BERTIN's fork of mesh-transformer-jax, which provides code adapt an Alpaca dataset to finetune any GPT-J-6B model. We run finetuning for 3 epochs using sequence length of 2048 on a single TPUv3-8 for 3 hours on top of BERTIN GPT-J-6B.
|
63 |
|
64 |
-
**
|
65 |
|
66 |
Esta es una versión GPTQ de 4 bits del [bertin-project/bertin-gpt-j-6B-alpaca]( https://huggingface.co/bertin-project/bertin-gpt-j-6B-alpaca)
|
67 |
|
|
|
16 |
|
17 |
<strong><span style="font-size: larger;">bertin-gpt-j-6B-alpaca-4bit-128g 🤗</span></strong>
|
18 |
|
19 |
+
![bertin](https://aeiljuispo.cloudimg.io/v7/https://s3.amazonaws.com/moonup/production/uploads/1627122830113-5ef3829e518622264685b0cd.png?w=200&h=200&f=face)
|
20 |
+
|
21 |
**descripción en español agregado ⬇️**
|
22 |
|
23 |
This is a 4-bit GPTQ version of the [bertin-project/bertin-gpt-j-6B-alpaca]( https://huggingface.co/bertin-project/bertin-gpt-j-6B-alpaca)
|
24 |
|
25 |
+
|
26 |
this is the result of quantizing to 4 bits using [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ).
|
27 |
|
28 |
** How to easily download and use this model in text-generation-webui** (tutorial by [TheBloke](https://huggingface.co/TheBloke))
|
|
|
64 |
Finetuning
|
65 |
To fine-tune the BERTIN GPT-J-6B model we used the code available on BERTIN's fork of mesh-transformer-jax, which provides code adapt an Alpaca dataset to finetune any GPT-J-6B model. We run finetuning for 3 epochs using sequence length of 2048 on a single TPUv3-8 for 3 hours on top of BERTIN GPT-J-6B.
|
66 |
|
67 |
+
**español**
|
68 |
|
69 |
Esta es una versión GPTQ de 4 bits del [bertin-project/bertin-gpt-j-6B-alpaca]( https://huggingface.co/bertin-project/bertin-gpt-j-6B-alpaca)
|
70 |
|