File size: 1,919 Bytes
8d46fb3
 
f10a05f
 
 
 
 
 
91f9006
8d46fb3
f10a05f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
---
license: mit
datasets:
- mwz/ur_para
language:
- ur
tags:
- 'paraphrase '
pipeline_tag: text2text-generation
---
# Urdu Paraphrasing Model

This repository contains a pretrained model for Urdu paraphrasing. The model is based on the BERT architecture and has been fine-tuned on a large dataset of Urdu paraphrases.

## Model Description

The pretrained model is based on the BERT architecture, specifically designed for paraphrasing tasks in the Urdu language. It has been trained using a large corpus of Urdu text to generate high-quality paraphrases.

## Model Details

- Model Name: Urdu-Paraphrasing-BERT
- Base Model: BERT
- Architecture: Transformer
- Language: Urdu
- Dataset: Urdu Paraphrasing Dataset mwz/ur_para

## How to Use

You can use this pretrained model for generating paraphrases for Urdu text. Here's an example of how to use the model:

```python
from transformers import pipeline

# Load the model
model = pipeline("text2text-generation", model="path_to_pretrained_model")

# Generate paraphrases
input_text = "Urdu input text for paraphrasing."
paraphrases = model(input_text, max_length=128, num_return_sequences=3)

# Print the generated paraphrases
print("Original Input Text:", input_text)
print("Generated Paraphrases:")
for paraphrase in paraphrases:
    print(paraphrase["generated_text"])
```
## Training
The model was trained using the Hugging Face transformers library. The training process involved fine-tuning the base BERT model on the Urdu Paraphrasing Dataset.

## Evaluation
The model's performance was evaluated on a separate validation set using metrics such as BLEU, ROUGE, and perplexity. However, please note that the evaluation results may vary depending on the specific use case.

## Acknowledgments
- The pretrained model is based on the BERT architecture developed by Google Research.

## License
This model and the associated code are licensed under the MIT License.