Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | @@ -49,6 +49,20 @@ This model was trained with DPO, a method introduced in [Direct Preference Optim | |
| 49 |  | 
| 50 | 
             
            ## Citations
         | 
| 51 |  | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 52 | 
             
            Cite DPO as:
         | 
| 53 |  | 
| 54 | 
             
            ```bibtex
         | 
|  | |
| 49 |  | 
| 50 | 
             
            ## Citations
         | 
| 51 |  | 
| 52 | 
            +
            Cite this model:
         | 
| 53 | 
            +
             | 
| 54 | 
            +
            ```bibtex
         | 
| 55 | 
            +
            @misc{lübbers2025enhancingparaphrasetypegeneration,
         | 
| 56 | 
            +
                  title={Enhancing Paraphrase Type Generation: The Impact of DPO and RLHF Evaluated with Human-Ranked Data}, 
         | 
| 57 | 
            +
                  author={Christopher Lee Lübbers},
         | 
| 58 | 
            +
                  year={2025},
         | 
| 59 | 
            +
                  eprint={2506.02018},
         | 
| 60 | 
            +
                  archivePrefix={arXiv},
         | 
| 61 | 
            +
                  primaryClass={cs.CL},
         | 
| 62 | 
            +
                  url={https://arxiv.org/abs/2506.02018}, 
         | 
| 63 | 
            +
            }
         | 
| 64 | 
            +
            ```
         | 
| 65 | 
            +
             | 
| 66 | 
             
            Cite DPO as:
         | 
| 67 |  | 
| 68 | 
             
            ```bibtex
         | 
