--- license: mit datasets: - mwz/ur_para language: - ur tags: - 'paraphrase ' pipeline_tag: text2text-generation --- # Urdu Paraphrasing Model This repository contains a pretrained model for Urdu paraphrasing. The model is based on the BERT architecture and has been fine-tuned on a large dataset of Urdu paraphrases. ## Model Description The pretrained model is based on the BERT architecture, specifically designed for paraphrasing tasks in the Urdu language. It has been trained using a large corpus of Urdu text to generate high-quality paraphrases. ## Model Details - Model Name: Urdu-Paraphrasing-BERT - Base Model: BERT - Architecture: Transformer - Language: Urdu - Dataset: Urdu Paraphrasing Dataset mwz/ur_para ## How to Use You can use this pretrained model for generating paraphrases for Urdu text. Here's an example of how to use the model: ```python from transformers import pipeline # Load the model model = pipeline("text2text-generation", model="path_to_pretrained_model") # Generate paraphrases input_text = "Urdu input text for paraphrasing." paraphrases = model(input_text, max_length=128, num_return_sequences=3) # Print the generated paraphrases print("Original Input Text:", input_text) print("Generated Paraphrases:") for paraphrase in paraphrases: print(paraphrase["generated_text"]) ``` ## Training The model was trained using the Hugging Face transformers library. The training process involved fine-tuning the base BERT model on the Urdu Paraphrasing Dataset. ## Evaluation The model's performance was evaluated on a separate validation set using metrics such as BLEU, ROUGE, and perplexity. However, please note that the evaluation results may vary depending on the specific use case. ## Acknowledgments - The pretrained model is based on the BERT architecture developed by Google Research. ## License This model and the associated code are licensed under the MIT License.