mwz commited on
Commit
404cef3
1 Parent(s): d1abd3a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -38
README.md CHANGED
@@ -6,52 +6,30 @@ language:
6
  - ur
7
  tags:
8
  - 'paraphrase '
9
- pipeline_tag: text2text-generation
10
  ---
11
- # Urdu Paraphrasing Model
12
 
13
- This repository contains a pretrained model for Urdu paraphrasing. The model is based on the BERT architecture and has been fine-tuned on a large dataset of Urdu paraphrases.
 
 
14
 
15
  ## Model Description
16
 
17
- The pretrained model is based on the BERT architecture, specifically designed for paraphrasing tasks in the Urdu language. It has been trained using a large corpus of Urdu text to generate high-quality paraphrases.
18
 
19
- ## Model Details
20
 
21
- - Model Name: Urdu-Paraphrasing-BERT
22
- - Base Model: BERT
23
- - Architecture: Transformer
24
- - Language: Urdu
25
- - Dataset: Urdu Paraphrasing Dataset mwz/ur_para
26
 
27
- ## How to Use
28
 
29
- You can use this pretrained model for generating paraphrases for Urdu text. Here's an example of how to use the model:
30
 
31
- ```python
32
- from transformers import pipeline
33
 
34
- # Load the model
35
- model = pipeline("text2text-generation", model="path_to_pretrained_model")
36
-
37
- # Generate paraphrases
38
- input_text = "Urdu input text for paraphrasing."
39
- paraphrases = model(input_text, max_length=128, num_return_sequences=3)
40
-
41
- # Print the generated paraphrases
42
- print("Original Input Text:", input_text)
43
- print("Generated Paraphrases:")
44
- for paraphrase in paraphrases:
45
- print(paraphrase["generated_text"])
46
- ```
47
- ## Training
48
- The model was trained using the Hugging Face transformers library. The training process involved fine-tuning the base BERT model on the Urdu Paraphrasing Dataset.
49
-
50
- ## Evaluation
51
- The model's performance was evaluated on a separate validation set using metrics such as BLEU, ROUGE, and perplexity. However, please note that the evaluation results may vary depending on the specific use case.
52
-
53
- ## Acknowledgments
54
- - The pretrained model is based on the BERT architecture developed by Google Research.
55
-
56
- ## License
57
- This model and the associated code are licensed under the MIT License.
 
6
  - ur
7
  tags:
8
  - 'paraphrase '
 
9
  ---
10
+ <h1 align="center">Urdu Paraphrase Generation Model</h1>
11
 
12
+ <p align="center">
13
+ <b>Fine-tuned model for Urdu paraphrase generation</b>
14
+ </p>
15
 
16
  ## Model Description
17
 
18
+ The Urdu Paraphrase Generation Model is a state-of-the-art language model trained on a large dataset of Urdu paraphrases. It is based on the `roberta-urdu-small` architecture, fine-tuned for the specific task of generating high-quality paraphrases.
19
 
20
+ ## Features
21
 
22
+ - Generate accurate and contextually relevant paraphrases in Urdu.
23
+ - Maintain linguistic nuances and syntactic structures of the original input.
24
+ - Handle a variety of input sentence lengths and complexities.
 
 
25
 
26
+ ## Usage
27
 
28
+ ### Installation
29
 
30
+ To use the Urdu Paraphrase Generation Model, follow these steps:
 
31
 
32
+ 1. Install the `transformers` library:
33
+ '''bash
34
+ pip install transformers
35
+ '''