Datasets:

Modalities:
Tabular
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas

Improve dataset card: Add metadata, paper link, project page, and sample usage

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +34 -1
README.md CHANGED
@@ -1,4 +1,13 @@
1
  ---
 
 
 
 
 
 
 
 
 
2
  dataset_info:
3
  features:
4
  - name: prompt
@@ -28,4 +37,28 @@ configs:
28
  path: data/train-*
29
  ---
30
 
31
- If you are the author of any comment in this dataset and would like it removed, please contact us and we will comply promptly.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ task_categories:
3
+ - text-ranking
4
+ tags:
5
+ - creative-writing
6
+ - llm-evaluation
7
+ - preference-alignment
8
+ - reward-modeling
9
+ - benchmark
10
+ - reddit
11
  dataset_info:
12
  features:
13
  - name: prompt
 
37
  path: data/train-*
38
  ---
39
 
40
+ # LitBench: A Benchmark and Dataset for Reliable Evaluation of Creative Writing
41
+
42
+ LitBench is the first standardized benchmark and paired dataset for reliable evaluation of creative writing generated by large language models (LLMs). It addresses the challenge of evaluating open-ended narratives, which lack ground truths. The dataset comprises a held-out test set of 2,480 debiased, human-labeled story comparisons drawn from Reddit and a 43,827-pair training corpus of human preference labels. LitBench facilitates benchmarking zero-shot LLM judges and training reward models for creative writing verification and optimization.
43
+
44
+ **Paper:** [LitBench: A Benchmark and Dataset for Reliable Evaluation of Creative Writing](https://huggingface.co/papers/2507.00769)
45
+
46
+ **Project Page (Hugging Face Collection):** https://huggingface.co/collections/SAA-Lab/litbench-68267b5da3aafe58f9e43461
47
+
48
+ ### Sample Usage
49
+
50
+ You can load the dataset using the Hugging Face `datasets` library:
51
+
52
+ ```python
53
+ from datasets import load_dataset
54
+
55
+ dataset = load_dataset("SAA-Lab/LitBench")
56
+
57
+ # Access the training split
58
+ train_dataset = dataset["train"]
59
+
60
+ # Print the first example
61
+ print(train_dataset[0])
62
+ ```
63
+
64
+ If you are the author of any comment in this dataset and would like it removed, please contact us and we will comply promptly.