Update README.md
Browse files
README.md
CHANGED
@@ -45,7 +45,7 @@ tokenizer = AutoTokenizer.from_pretrained("urduhack/roberta-urdu-small")
|
|
45 |
Use the following code snippet to generate paraphrases with the loaded model and tokenizer:
|
46 |
```
|
47 |
# Example sentence
|
48 |
-
input_sentence = "
|
49 |
|
50 |
# Tokenize the input sentence
|
51 |
inputs = tokenizer(input_sentence, truncation=True, padding=True, return_tensors="pt")
|
@@ -59,3 +59,13 @@ with torch.no_grad():
|
|
59 |
paraphrase = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
60 |
print("Paraphrase:", paraphrase)
|
61 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
Use the following code snippet to generate paraphrases with the loaded model and tokenizer:
|
46 |
```
|
47 |
# Example sentence
|
48 |
+
input_sentence = "تصوراتی طور پر کریم سکمنگ کی دو بنیادی جہتیں ہیں - مصنوعات اور جغرافیہ۔"
|
49 |
|
50 |
# Tokenize the input sentence
|
51 |
inputs = tokenizer(input_sentence, truncation=True, padding=True, return_tensors="pt")
|
|
|
59 |
paraphrase = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
60 |
print("Paraphrase:", paraphrase)
|
61 |
```
|
62 |
+
|
63 |
+
## Performance
|
64 |
+
The model has been fine-tuned on a 30k rows dataset of Urdu paraphrases and achieves impressive performance in generating high-quality paraphrases. Detailed performance metrics, such as accuracy and fluency, are being evaluated and will be updated soon.
|
65 |
+
|
66 |
+
## Contributing
|
67 |
+
Contributions to the Urdu Paraphrase Generation Model are welcome! If you find any issues or have suggestions for improvements, please open an issue or submit a pull request.
|
68 |
+
|
69 |
+
## License
|
70 |
+
|
71 |
+
This project is licensed under the MIT License. See the [LICENSE](LICENSE) file for details.
|