nielsr HF Staff commited on
Commit
7879e60
·
verified ·
1 Parent(s): 2512923

Improve model card: Add Transformers library tag, paper link, GitHub link, and usage details

Browse files

This PR enhances the model card for the MahaParaphrase-BERT model by:

- Adding `library_name: transformers` to the metadata, which enables proper integration and displays an auto-generated usage snippet on the model page, as the model's configuration indicates compatibility with the `transformers` library.
- Updating the paper link to the official publication: [MahaParaphrase: A Marathi Paraphrase Detection Corpus and BERT-based Models](https://huggingface.co/papers/2508.17444).
- Adding an explicit link to the [L3Cube-MahaNLP GitHub repository](https://github.com/l3cube-pune/MarathiNLP) for code and resources.
- Including a "Usage" section that directs users to the GitHub repository and an associated Colab notebook for practical examples.
- Updating the citation section to refer to the specific paper for this model.

Please review and merge this PR if these improvements are acceptable.

Files changed (1) hide show
  1. README.md +19 -11
README.md CHANGED
@@ -1,30 +1,38 @@
1
  ---
2
- license: cc-by-4.0
3
  datasets:
4
  - l3cube-pune/MahaParaphrase
5
  language:
6
  - mr
 
7
  metrics:
8
  - accuracy
9
  pipeline_tag: text-classification
 
10
  ---
 
11
  # MahaParaphrase-BERT Model
12
 
13
  **MahaParaphrase-BERT** is a **MahaBERT** model ([l3cube-pune/marathi-bert-v2](https://huggingface.co/l3cube-pune/marathi-bert-v2)) fine-tuned on the **L3Cube-MahaParaphrase Dataset**, a high-quality Marathi paraphrase detection corpus. The dataset consists of 8,000 sentence pairs annotated as Paraphrase (P) or Non-paraphrase (NP).
14
 
15
  This model is trained specifically for Marathi Paraphrase Detection task.
16
 
17
- For more details on the **L3Cube-MahaParaphrase Dataset**, visit the [dataset link](https://github.com/l3cube-pune/MarathiNLP).
 
 
 
 
 
 
18
 
19
- More details on the model, training methodology, and evaluation results can be found in our [paper](coming soon).
20
 
21
- ## Citing:
22
- If you use this model, please cite it as follows:
23
- ```
24
- @article{joshi2022l3cube,
25
- title={L3cube-mahanlp: Marathi natural language processing datasets, models, and library},
26
- author={Joshi, Raviraj},
27
- journal={arXiv preprint arXiv:2205.14728},
28
- year={2022}
29
  }
30
  ```
 
1
  ---
 
2
  datasets:
3
  - l3cube-pune/MahaParaphrase
4
  language:
5
  - mr
6
+ license: cc-by-4.0
7
  metrics:
8
  - accuracy
9
  pipeline_tag: text-classification
10
+ library_name: transformers
11
  ---
12
+
13
  # MahaParaphrase-BERT Model
14
 
15
  **MahaParaphrase-BERT** is a **MahaBERT** model ([l3cube-pune/marathi-bert-v2](https://huggingface.co/l3cube-pune/marathi-bert-v2)) fine-tuned on the **L3Cube-MahaParaphrase Dataset**, a high-quality Marathi paraphrase detection corpus. The dataset consists of 8,000 sentence pairs annotated as Paraphrase (P) or Non-paraphrase (NP).
16
 
17
  This model is trained specifically for Marathi Paraphrase Detection task.
18
 
19
+ More details on the model, training methodology, and evaluation results can be found in the paper: [MahaParaphrase: A Marathi Paraphrase Detection Corpus and BERT-based Models](https://huggingface.co/papers/2508.17444).
20
+
21
+ For further resources, including the **L3Cube-MahaParaphrase Dataset** and code, visit the [L3Cube-MahaNLP GitHub repository](https://github.com/l3cube-pune/MarathiNLP).
22
+
23
+ ## Usage
24
+
25
+ This model can be used with the Hugging Face `transformers` library. For detailed usage examples and more information on the `mahaNLP` library and its integration, please refer to the [L3Cube-MahaNLP GitHub repository](https://github.com/l3cube-pune/MarathiNLP) and its associated [Colab notebook](https://colab.research.google.com/drive/1POx3Bi1cML6-s3Z3u8g8VpqzpoYCyv2q).
26
 
27
+ ## Citing
28
 
29
+ If you use this model or the associated dataset, please cite the following paper:
30
+ ```bibtex
31
+ @article{mahaparaphrase2025,
32
+ title={MahaParaphrase: A Marathi Paraphrase Detection Corpus and BERT-based Models},
33
+ author={Please refer to the paper for authors},
34
+ journal={arXiv preprint arXiv:2508.17444},
35
+ year={2025},
36
+ url={https://arxiv.org/abs/2508.17444}
37
  }
38
  ```