Model Card for Model ID

RaDeR, are a set of reasoning-based dense retrieval and reranker models trained with data derived from mathematical problem solving using large language models (LLMs). RaDeR retrievers, trained for mathematical reasoning, effectively generalize to diverse retrieval reasoning tasks in the BRIGHT and RAR-b benchmarks, consistently outperforming strong baselines in overall performance.

Model Details

Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Developed by: CIIR, UMass Amherst
  • Model type: Retriever
  • Language(s): English
  • License: MIT
  • Finetuned from model: Alibaba-NLP/gte-Qwen2-7B-instruct

Model Sources

How to Get Started with the Model

Run the following code to start a server of the model with vLLM for fast inference.

vllm serve Raderspace/RaDeR_gte_Qwen2-7B_MATH_LLMq_CoT_lexical \
  --task embed \
  --trust-remote-code \
  --override-pooler-config '{"pooling_type": "LAST", "normalize": true}' \
  --gpu-memory-utilization 0.9 \
  --api-key abc \
  --tokenizer Alibaba-NLP/gte-Qwen2-7B-instruct \
  --port 8001 \
  --disable-log-requests \
  --max-num-seqs 5000

Follow the code on Github to see how to query the retriever server.

Training Details

Training Data

The model was trained using the MATH retrieval training dataset from RaDeR, containing CoT, LLMq and lexical query types.

Software

https://github.com/Debrup-61/RaDeR

Citation [optional]

BibTeX:

@misc{das2025raderreasoningawaredenseretrieval,
      title={RaDeR: Reasoning-aware Dense Retrieval Models}, 
      author={Debrup Das and Sam O' Nuallain and Razieh Rahimi},
      year={2025},
      eprint={2505.18405},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2505.18405}, 
}

Model Card Contact

Debrup Das: [email protected]

Downloads last month
2
Safetensors
Model size
7.07B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Raderspace/RaDeR_gte_Qwen2-7B_MATH_LLMq_CoT_lexical

Finetuned
(8)
this model

Dataset used to train Raderspace/RaDeR_gte_Qwen2-7B_MATH_LLMq_CoT_lexical

Collection including Raderspace/RaDeR_gte_Qwen2-7B_MATH_LLMq_CoT_lexical