Update README.md
Browse files
README.md
CHANGED
@@ -14,7 +14,7 @@ widget:
|
|
14 |
|
15 |
inference:
|
16 |
parameters:
|
17 |
-
max_length:
|
18 |
---
|
19 |
|
20 |
# GPT-2
|
@@ -31,13 +31,9 @@ model = GPTNeoForCausalLM.from_pretrained('minhtoan/gpt3-small-vietnamese')
|
|
31 |
|
32 |
text = "Hoa quả và rau thường rẻ hơn khi vào mùa"
|
33 |
input_ids = tokenizer.encode(text, return_tensors='pt')
|
34 |
-
max_length =
|
35 |
|
36 |
-
sample_outputs = model.generate(input_ids,
|
37 |
-
do_sample=True,
|
38 |
-
max_length=max_length,
|
39 |
-
min_length=max_length,
|
40 |
-
num_return_sequences=1)
|
41 |
|
42 |
for i, sample_output in enumerate(sample_outputs):
|
43 |
print(">> Generated text {}\n\n{}".format(i+1, tokenizer.decode(sample_output.tolist())))
|
|
|
14 |
|
15 |
inference:
|
16 |
parameters:
|
17 |
+
max_length: 80
|
18 |
---
|
19 |
|
20 |
# GPT-2
|
|
|
31 |
|
32 |
text = "Hoa quả và rau thường rẻ hơn khi vào mùa"
|
33 |
input_ids = tokenizer.encode(text, return_tensors='pt')
|
34 |
+
max_length = 80
|
35 |
|
36 |
+
sample_outputs = model.generate(input_ids, do_sample=True, max_length=max_length)
|
|
|
|
|
|
|
|
|
37 |
|
38 |
for i, sample_output in enumerate(sample_outputs):
|
39 |
print(">> Generated text {}\n\n{}".format(i+1, tokenizer.decode(sample_output.tolist())))
|