Datasets:

Modalities:
Text
Formats:
json
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
pandas
License:
Dataset Viewer
Auto-converted to Parquet
query-id
stringlengths
2
4
corpus-id
stringlengths
2
4
score
float64
1
1
q15
c20
1
q81
c23
1
q80
c74
1
q16
c54
1
q110
c87
1
q90
c62
1
q27
c64
1
q28
c9
1
q83
c36
1
q84
c65
1
q53
c102
1
q34
c98
1
q79
c32
1
q64
c8
1
q63
c93
1
q47
c107
1
q25
c57
1
q24
c35
1
q72
c17
1
q38
c49
1
q35
c115
1
q92
c85
1
q93
c97
1
q95
c12
1
q94
c101
1
q91
c60
1
q105
c58
1
q104
c86
1
q50
c22
1
q13
c90
1
q12
c21
1
q101
c19
1
q98
c5
1
q99
c16
1
q100
c105
1
q52
c25
1
q51
c111
1
q10
c94
1
q11
c114
1
q55
c82
1
q22
c67
1
q6
c103
1
q7
c43
1
q8
c6
1
q114
c27
1
q113
c55
1
q111
c18
1
q112
c83
1
q106
c108
1
q75
c52
1
q68
c53
1
q88
c95
1
q3
c4
1
q69
c88
1
q42
c15
1
q86
c69
1
q109
c78
1
q39
c112
1
q77
c29
1
q82
c28
1
q116
c2
1
q46
c24
1
q18
c91
1
q60
c76
1
q96
c40
1
q103
c50
1
q73
c68
1
q33
c30
1
q71
c45
1
q32
c39
1
q45
c89
1
q44
c70
1
q40
c84
1
q61
c11
1
q62
c48
1
q66
c3
1
q65
c41
1
q76
c73
1
q59
c44
1
q85
c66
1
q54
c92
1
q4
c104
1
q87
c38
1
q30
c51
1
q5
c13
1
q1
c96
1
q2
c99
1
q31
c46
1
q17
c59
1
q37
c109
1
q36
c81
1
q26
c26
1
q49
c31
1
q56
c71
1
q48
c56
1
q43
c72
1
q78
c30
1
q102
c113
1
q14
c7
1
q57
c42
1
End of preview. Expand in Data Studio

Bar Exam QA MTEB Benchmark πŸ‹

This is the test split of the Bar Exam QA dataset formatted in the Massive Text Embedding Benchmark (MTEB) information retrieval dataset format.

This dataset is intended to facilitate the consistent and reproducible evaluation of information retrieval models on Bar Exam QA with the mteb embedding model evaluation framework.

More specifically, this dataset tests the ability of information retrieval models to identify legal provisions relevant to US bar exam questions.

This dataset has been processed into the MTEB format by Isaacus, a legal AI research company.

Methodology πŸ§ͺ

To understand how Bar Exam QA was created, refer to its documentation.

This dataset was formatted by concatenating the prompt and question columns of the source data delimited by a single space (or, where there was no prompt, reverting to the question only) into queries (or anchors), and treating the gold_passage column as relevant (or positive) passages.

Structure πŸ—‚οΈ

As per the MTEB information retrieval dataset format, this dataset comprises three splits, default, corpus and queries.

The default split pairs queries (query-id) with relevant passages (corpus-id), each pair having a score of 1.

The corpus split contains relevant passages from Bar Exam QA, with the text of a passage being stored in the text key and its id being stored in the _id key.

The queries split contains queries, with the text of a query being stored in the text key and its id being stored in the _id key.

License πŸ“œ

To the extent that any intellectual property rights reside in the contributions made by Isaacus in formatting and processing this dataset, Isaacus licenses those contributions under the same license terms as the source dataset. You are free to use this dataset without citing Isaacus.

The source dataset is licensed under CC BY SA 4.0.

Citation πŸ”–

@inproceedings{Zheng_2025, series={CSLAW ’25},
   title={A Reasoning-Focused Legal Retrieval Benchmark},
   url={http://dx.doi.org/10.1145/3709025.3712219},
   DOI={10.1145/3709025.3712219},
   booktitle={Proceedings of the Symposium on Computer Science and Law on ZZZ},
   publisher={ACM},
   author={Zheng, Lucia and Guha, Neel and Arifov, Javokhir and Zhang, Sarah and Skreta, Michal and Manning, Christopher D. and Henderson, Peter and Ho, Daniel E.},
   year={2025},
   month=mar, pages={169–193},
   collection={CSLAW ’25},
   eprint={2505.03970}
}
Downloads last month
162