Dataset Viewer
instance_id
int64 1
31
| paper_name
stringlengths 10
101
| folder_name
stringlengths 5
26
| paper_url
stringlengths 32
50
| year
int64 2.02k
2.03k
| repo_url
stringlengths 30
90
| repo_folder_name
stringlengths 3
31
| implementations
listlengths 1
5
|
---|---|---|---|---|---|---|---|
11 |
MoverScore: Text Generation Evaluating with Contextualized Embeddings and Earth Mover Distance
|
11-moverscore
|
https://aclanthology.org/D19-1053.pdf
| 2,019 |
https://github.com/AIPHES/emnlp19-moverscore
|
moverscore
|
[
{
"category": "Evaluation Metrics & Benchmarking",
"class_name": "",
"goal_file": "movescore.py",
"goal_function": "word_mover_score",
"golden_file": "golden_files/movescore_golden.py",
"index": 1,
"instruction": "You need to implement the word_mover_score function in movescore.py based on paper.pdf, which calculates a similarity score between two sentences by using BERT embeddings and by calculating calculating the value of 1-Word Mover's Distance metric",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
13 |
Probing the Decision Boundaries of In-context Learning in Large Language Models
|
13-DecBound
|
https://arxiv.org/pdf/2406.11233
| 2,024 |
https://github.com/siyan-zhao/ICL_decision_boundary
|
DecBound-main
|
[
{
"category": "Data Augmentation & Generation",
"class_name": "",
"goal_file": "data_utils.py",
"goal_function": "generate_tasks",
"golden_file": "golden_files/data_utils_golden.py",
"index": 1,
"instruction": "Implement `generate_tasks` function in data_utils.py based on paper.pdf and the repository, which generates multiple machine learning tasks, each represented as a separate dataset.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test.py"
}
] |
14 |
EmojiPrompt: Generative Prompt Obfuscation for Privacy-Preserving Communication with Cloud-based LLMs
|
14-EmojiCrypt
|
https://arxiv.org/pdf/2402.05868
| 2,025 |
https://github.com/agiresearch/EmojiCrypt
|
EmojiCrypt-main
|
[
{
"category": "Evaluation Metrics",
"class_name": "",
"goal_file": "Tabular.py",
"goal_function": "compute_mean_cosine_similarity",
"golden_file": "golden_files/Tabular_golden.py",
"index": 1,
"instruction": "Implement `compute_mean_cosine_similarity` function in Tabular.py based on the paper.pdf and the repository. Computes the mean cosine similarity between pairs of encrypted and decrypted feature levels.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
15 |
Language Models Predict Empathy Gaps Between Social In-groups and Out-groups
|
15-EmpathyBias
|
https://arxiv.org/pdf/2503.01030
| 2,025 |
https://github.com/houyu0930/intergroup-empathy-bias
|
EmpathyBias-main
|
[
{
"category": "Evaluation Metrics",
"class_name": "",
"goal_file": "analysis/analysis_after_processing.py",
"goal_function": "get_delta",
"golden_file": "golden_files/analysis_after_processing_golden.py",
"index": 1,
"instruction": "Implement `get_delta` function analysis/analysis_after_processing.py based on paper.pdf and the repository. Compute the difference in average intensity between in-group and out-group interactions.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
},
{
"category": "Evaluation Metrics",
"class_name": "",
"goal_file": "analysis/analysis_after_processing.py",
"goal_function": "get_filtered_matrix",
"golden_file": "golden_files/analysis_after_processing_golden.py",
"index": 2,
"instruction": "Implement `get_filtered_matrix` function analysis/analysis_after_processing.py based on paper.pdf and the repository. Compute the standardized emotion intensity matrix for a specific group, model, and prompt variation, excluding specified data points (by ID).",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_2.py"
}
] |
16 |
Probing Language Models on Their Knowledge Source
|
16-KnowProb
|
https://aclanthology.org/2024.blackboxnlp-1.35.pdf
| 2,024 |
https://github.com/Zineddine-Tighidet/knowledge-probing-framework
|
KnowProb-main
|
[
{
"category": "Interpretability & Explainability",
"class_name": "",
"goal_file": "src/classification/classifier.py",
"goal_function": "perform_classification_by_relation_group",
"golden_file": "golden_files/classifier_golden.py",
"index": 1,
"instruction": "Implement `perform_classification_by_relation_group` function in src/classification/classifier.py based on paper.pdf and the repository. This function performs classification over the saved activations, the classification can be module-wise and token-wise (i.e. vertical).",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
},
{
"category": "Data Augmentation & Generation",
"class_name": "",
"goal_file": "src/data.py",
"goal_function": "remove_object_subject_overlap",
"golden_file": "golden_files/data_golden.py",
"index": 2,
"instruction": "Implement `remove_object_subject_overlap` function in src/data.py based on paper.pdf and the repository. This function deletes the examples where the subject and the parametric object are similar using a Jaro-Winkler string similarity.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_2.py"
},
{
"category": "Information Extraction & Knowledge Integration",
"class_name": "",
"goal_file": "src/parametric_knowledge.py",
"goal_function": "is_parametric_object_not_in_the_prompt",
"golden_file": "golden_files/parametric_knowledge_golden.py",
"index": 3,
"instruction": "Implement `is_parametric_object_not_in_the_prompt` function in src/parametric_knowledge.py based on paper.pdf and the repository. This function checks whether the parametric_object is included in the one-shot examples that were used to guide the LLM during parametric knowledge building as it would mean that it's biased by the prompt.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_3.py"
},
{
"category": "Information Extraction & Knowledge Integration",
"class_name": "",
"goal_file": "src/data.py",
"goal_function": "return_entity_overlap_between_relation_groups",
"golden_file": "golden_files/data_golden.py",
"index": 4,
"instruction": "Implement `return_entity_overlap_between_relation_groups` function in src/data.py based on paper.pdf and the repository. This function returns a dict of dicts representing a kind of matrix of shape (nb_relation_groups, nb_relation_groups), where each value is the number of examples in a relation group that have the same subject or object in another relation group.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_4.py"
},
{
"category": "Data Augmentation & Generation",
"class_name": "",
"goal_file": "src/data.py",
"goal_function": "generate_counter_parametric_knowledge_dataset",
"golden_file": "golden_files/data_golden.py",
"index": 5,
"instruction": "Implement `generate_counter_parametric_knowledge_dataset` function in src/data.py based on paper.pdf and the repository. This function generates the counter-parametric knowledge using the parametric knowledge.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_5.py"
}
] |
18 |
Massive Activations in Large Language Models
|
18-MassActiv
|
https://arxiv.org/abs/2402.17762
| 2,024 |
https://github.com/locuslab/massive-activations
|
MassActiv-main
|
[
{
"category": "Interpretability & Explainability",
"class_name": "",
"goal_file": "main_vit.py",
"goal_function": "run_exp1",
"golden_file": "golden_files/main_vit_golden.py",
"index": 1,
"instruction": "Implement run_exp1 function in main_vit.py. Run 3D feature visualization for a specific layer in a Vision Transformer (ViT) model. This function extracts the absolute feature activations from a specified transformer layer, visualizes them in 3D, and saves both the input parameters and feature activations as serialized pickle files for reproducibility and analysis.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test.py"
}
] |
19 |
Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention
|
19-Native-Sparse-Attention
|
https://arxiv.org/abs/2502.11089
| 2,025 |
https://github.com/fla-org/native-sparse-attention
|
NSA
|
[
{
"category": "Neural Network Architectures & Modules",
"class_name": "",
"goal_file": "NSA.py",
"goal_function": "nsa",
"golden_file": "golden_files/NSA_golden.py",
"index": 1,
"instruction": "Please implement the `nsa` function in NSA.py based on the Natively trainable Sparse Attention mechanism mentioned in the paper.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
20 |
Trusting Your Evidence: Hallucinate Less with Context-aware Decoding
|
20-CAD
|
https://arxiv.org/abs/2305.14739
| 2,023 |
https://github.com/xhan77/context-aware-decoding.git
|
context-aware-decoding-main
|
[
{
"category": "Decoding & Search Strategies",
"class_name": "",
"goal_file": "unit_test/unit_test_1.py",
"goal_function": "context_aware_sampling",
"golden_file": "golden_files/unit_test_1_golden.py",
"index": 1,
"instruction": "Implement the function context_aware_sampling in the unit_test/unit_test_1.py and pass the test by running unit_test/unit_test_1.py file, you may ignore the following parameters: alpha, max_length, temperature",
"retrieval_content": null,
"retrieval_context": [],
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
22 |
Token-level Direct Preference Optimization
|
22-TokenDPO
|
https://arxiv.org/abs/2404.11999
| 2,024 |
https://github.com/Vance0124/Token-level-Direct-Preference-Optimization/tree/master
|
TokenDPO-main
|
[
{
"category": "Training Objectives & Optimization Techniques",
"class_name": "",
"goal_file": "trainers.py",
"goal_function": "tdpo_loss",
"golden_file": "golden_files/trainers_golden.py",
"index": 1,
"instruction": "Implement the `tdpo_loss` function in trainers.py based on the Token-level DPO loss mentioned in paper.pdf and the code repository. You may ignore the following parameters: if_tdpo2.",
"retrieval_content": null,
"retrieval_context": [],
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
23 |
A Simple Framework for Contrastive Learning of Visual Representations
|
23-SimCLR
|
https://arxiv.org/pdf/2002.05709
| 2,020 |
https://github.com/skywalkerzhang/SimCLR/tree/master
|
SimCLR-main
|
[
{
"category": "Training Objectives & Optimization Techniques",
"class_name": "SimCLR",
"goal_file": "simclr.py",
"goal_function": "info_nce_loss",
"golden_file": "golden_files/simclr_golden.py",
"index": 1,
"instruction": "Implement the `info_nce_loss` method of `SimCLR` class in simclr.py that computes the InfoNCE loss for a batch of features, based on the method described in the paper.pdf and the repository.",
"retrieval_content": null,
"retrieval_context": [],
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
25 |
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning
|
25-AdaLora
|
https://arxiv.org/pdf/2303.10512
| 2,023 |
https://github.com/QingruZhang/AdaLoRA/tree/main
|
AdaLoRA
|
[
{
"category": "Neural Network Architectures & Modules",
"class_name": "SVDLinear",
"goal_file": "loralib/loralib/adalora.py",
"goal_function": "forward",
"golden_file": "golden_files/adalora_golden.py",
"index": 1,
"instruction": "Implement the forward function of SVD-based Adaptated linear layer in loralib/loralib/adalora.py mentioned in the section 3.1 of the paper.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
28 |
Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models
|
28-SPIN
|
https://arxiv.org/abs/2401.01335
| 2,024 |
https://github.com/uclaml/SPIN
|
SPIN
|
[
{
"category": "loss function",
"class_name": "SPINTrainer",
"goal_file": "spin/alignment/trainer.py",
"goal_function": "spin_loss",
"golden_file": "golden_files/trainer_golden.py",
"index": 1,
"instruction": "Implement the `spin_loss` method of `SPINTrainer` class in `.spin/alignment/trainer.py` based on the loss function of SPIN algorithm mentioned in the paper.pdf.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
30 |
ERGO: Event Relational Graph Transformer for Document-level Event Causality Identification
|
30-ERGO
|
https://aclanthology.org/2022.coling-1.185.pdf
| 2,022 |
https://github.com/chenmeiqii/ERGO
|
ERGO
|
[
{
"category": "Training Objectives & Optimization Techniques",
"class_name": "focal_loss",
"goal_file": "model.py",
"goal_function": "forward",
"golden_file": "golden_files/model_golden.py",
"index": 1,
"instruction": "Implement the loss function in the focal_loss class in model.py based on the final adaptive focal loss mentioned in the paper and the code repository. Justify whether summing or averaging the loss based on the `self.size_average` value.",
"retrieval_content": null,
"retrieval_context": [],
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
31 |
Aspect-Based Sentiment Analysis with Syntax-Opinion-Sentiment Reasoning Chain
|
31-Syn-Chain
|
https://aclanthology.org/2025.coling-main.210.pdf
| 2,025 |
https://github.com/rf-x/Syn-Chain-ABSA
|
Syn-Chain-ABSA
|
[
{
"category": "",
"class_name": "",
"goal_file": "conll_tree.py",
"goal_function": "spacy_result_to_conll",
"golden_file": "golden_files/conll_tree_golden.py",
"index": 1,
"instruction": "Implement the spacy_result_to_conll function in conll_tree.py to generate a dependency parse in the CoNLL-U format, as mentioned in the paper and the code repository.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
1 |
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
|
01-DPO
|
https://arxiv.org/pdf/2305.18290
| 2,023 |
https://github.com/eric-mitchell/direct-preference-optimization
|
direct-preference-optimization
|
[
{
"category": "Training Objectives & Optimization Techniques",
"class_name": "",
"goal_file": "trainers.py",
"goal_function": "preference_loss",
"golden_file": "golden_files/trainers_golden.py",
"index": 1,
"instruction": "Implement the preference_loss function in trainers.py based on the DPO loss mentioned in the paper and the code repository. You may ignore the following parameters: ipo, reference_free and label_smoothing.",
"retrieval_content": null,
"retrieval_context": [],
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
2 |
Language Models as Hierarchy Encoders
|
02-HITs
|
https://arxiv.org/pdf/2401.11374
| 2,024 |
https://github.com/KRR-Oxford/HierarchyTransformers
|
HierarchyTransformers
|
[
{
"category": "Training Objectives & Optimization Techniques",
"class_name": "HierarchyTransformerLoss",
"goal_file": "src/hierarchy_transformers/losses/hit_loss.py",
"goal_function": "forward",
"golden_file": "golden_files/hit_loss_golden.py",
"index": 1,
"instruction": "Implement the forward function of calulating the Hyperbolic Loss in src/hierarchy_transformers/losses/hit_loss.py based on paper.pdf.",
"retrieval_content": null,
"retrieval_context": [
"src/hierarchy_transformers/models/hierarchy_transformer/hit.py",
"src/hierarchy_transformers/models/hierarchy_transformer/hyperbolic.py"
],
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
3 |
DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models
|
03-DoLa
|
https://arxiv.org/abs/2309.03883
| 2,023 |
https://github.com/voidism/DoLa.git
|
DoLa-main
|
[
{
"category": "Decoding & Search Strategies",
"class_name": "",
"goal_file": "unit_test/unit_test_1.py",
"goal_function": "dola_greedy_decode_agent",
"golden_file": "golden_files/unit_test_1_golden.py",
"index": 1,
"instruction": "Implement the function dola_greedy_decode_agent in the unit_test/unit_test_1.py and pass the test by running unit_test/unit_test_1.py file",
"retrieval_content": null,
"retrieval_context": [],
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
4 |
Exploring Concept Depth: How Large Language Models Acquire Knowledge and Concept at Different Layers?
|
04-CD
|
https://arxiv.org/abs/2404.07066
| 2,024 |
https://github.com/Luckfort/CD
|
CD-main
|
[
{
"category": "Interpretability & Explainability",
"class_name": "",
"goal_file": "main.py",
"goal_function": "probing",
"golden_file": "golden_files/main_golden.py",
"index": 1,
"instruction": "Implement `probing` function in main.py based on paerp.pdf and the repository. Generate Probing function to evaluate whether the model can effectively distinguish between different data based on hidden states from different layers, using logistic regression classifier.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
5 |
SimPO: Simple Preference Optimization with a Reference-Free Reward
|
05-SimPO
|
https://arxiv.org/abs/2405.14734
| 2,024 |
https://github.com/princeton-nlp/SimPO
|
SimPO
|
[
{
"category": "loss function",
"class_name": "SimPOTrainer",
"goal_file": "scripts/simpo_trainer.py",
"goal_function": "simpo_loss",
"golden_file": "golden_files/simpo_trainer_golden.py",
"index": 1,
"instruction": "Implement the `simpo_loss` function of `SimPOTrainer` class in './scripts/simpo_trainer.py' based on the SimPO loss descripted in the paper.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
6 |
Representation Engineering: A Top-Down Approach to AI Transparency
|
06-RepE
|
https://arxiv.org/abs/2310.01405
| 2,023 |
https://github.com/andyzoujm/representation-engineering
|
representation-engineering-main
|
[
{
"category": "Feature Learning & Representation",
"class_name": "",
"goal_file": "unit_test/unit_test_1.py",
"goal_function": "get_rep_directions_agent",
"golden_file": "golden_files/unit_test_1_golden.py",
"index": 1,
"instruction": "Implement the function get_rep_directions_agent in the unit_test/unit_test_1.py and pass the test by running unit_test/unit_test_1.py file",
"retrieval_content": null,
"retrieval_context": [],
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
7 |
BERTScore: Evaluating Text Generation with BERT
|
07-bertscore
|
https://openreview.net/pdf?id=SkeHuCVFDr
| 2,020 |
https://github.com/Tiiiger/bert_score/tree/master
|
bert_score
|
[
{
"category": "Evaluation Metrics",
"class_name": "",
"goal_file": "bert_score_goal.py",
"goal_function": "greedy_cos_idf",
"golden_file": "golden_files/bert_score_golden.py",
"index": 1,
"instruction": "You need to implement the greedy_cos_idf function in bert_score_goal.py based on paper.pdf, which calculates a similarity score between two sentences by using BERT embeddings and performing a greedy matching between all reference and candidate words, with cosine similarity between vector representations as the scoring",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
8 |
StepLength
|
08-StepLength
|
https://arxiv.org/abs/2401.04925
| 2,024 |
https://github.com/MingyuJ666/The-Impact-of-Reasoning-Step-Length-on-Large-Language-Models
|
stepLength-main
|
[
{
"category": "Prompt Engineering & Instruction Tuning",
"class_name": "",
"goal_file": "run_inference.py",
"goal_function": "get_sentence",
"golden_file": "golden_files/run_inference_golden.py",
"index": 1,
"instruction": "Implement `get_sentence` function in run_inference.py based on paper.pdf and the repository. This code constructs the input k for an agent based on the specified prompting method args_method. If the method is 'zero_shot', it appends a direct answer trigger to the question x. If it is 'zero_shot_cot', it appends a chain-of-thought trigger to x. For 'few_shot' and 'few_shot_cot', it prepends the demonstration demo1 to x. If the method is 'auto_cot', it combines demo1, x, and the CoT trigger. The final string k is used as the input for the agent to simulate how the agent would behave under different prompting strtegies.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
9 |
Prompt-Based Monte-Carlo Tree Search for Goal-oriented Dialogue Policy Planning
|
09-GDPZero
|
https://aclanthology.org/2023.emnlp-main.439.pdf
| 2,023 |
https://github.com/jasonyux/GDPZero
|
GDPZero
|
[
{
"category": "Decoding & Search Strategies",
"class_name": "OpenLoopMCTS",
"goal_file": "core/mcts.py",
"goal_function": "find_best_action",
"golden_file": "golden_files/mcts_golden.py",
"index": 1,
"instruction": "Implement the self.find_best_action function in the OpenLoopMCTS class in core/mcts.py based on the PUCT formula mentioned in the paper and the code repository.",
"retrieval_content": [],
"retrieval_context": null,
"unit_test_file": "unit_test/unit_test_1.py"
}
] |
README.md exists but content is empty.
- Downloads last month
- 24