Datasets:

Modalities:
Text
Formats:
parquet
Size:
< 1K
ArXiv:
Libraries:
Datasets
pandas
jiang719 commited on
Commit
f2c607e
·
verified ·
1 Parent(s): 0352d61

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -4
README.md CHANGED
@@ -282,6 +282,7 @@ configs:
282
  - split: test
283
  path: data/test-*
284
  ---
 
285
  # Can Language Models Replace Programmers? REPOCOD Says 'Not Yet'
286
 
287
  Large language models (LLMs) have achieved high accuracy, i.e., more than 90 pass@1, in solving Python coding problems in HumanEval and MBPP. Thus, a natural question is, whether LLMs achieve comparable code completion performance compared to human developers? Unfortunately, one cannot answer this question using existing manual crafted or simple (e.g., single-line) code generation benchmarks, since such tasks fail to represent real-world software development tasks. In addition, existing benchmarks often use poor code correctness metrics, providing misleading conclusions.
@@ -298,9 +299,8 @@ REPOCOD_Lite_Unified is a variation of REPOCOD-Lite that has a similar format as
298
 
299
  * Check our [Leaderboard](https://lt-asset.github.io/REPOCOD/) for preliminary results using SOTA LLMs with RAG.
300
 
301
- *
302
- ```
303
 
 
304
  "instance_id": Instance ID in REPOCOD
305
  "version": Version of REPOCOD
306
  "gold_patches": {
@@ -322,13 +322,11 @@ REPOCOD_Lite_Unified is a variation of REPOCOD-Lite that has a similar format as
322
  "code": Problem statement for code generation.
323
  "test": Problem statement for test generation.
324
  }
325
- # "problem_statement_source": "repocod",
326
  "environment_setup_commit": base commit
327
  "evaluation": {
328
  "FAIL_TO_PASS": list of relevant test cases
329
  "PASS_TO_PASS": None, (all remaining tests that passes, we choose not to run the PASS_TO_PASS tests to avoid the computational cost)
330
  }
331
-
332
  ```
333
 
334
  ## Citation
 
282
  - split: test
283
  path: data/test-*
284
  ---
285
+
286
  # Can Language Models Replace Programmers? REPOCOD Says 'Not Yet'
287
 
288
  Large language models (LLMs) have achieved high accuracy, i.e., more than 90 pass@1, in solving Python coding problems in HumanEval and MBPP. Thus, a natural question is, whether LLMs achieve comparable code completion performance compared to human developers? Unfortunately, one cannot answer this question using existing manual crafted or simple (e.g., single-line) code generation benchmarks, since such tasks fail to represent real-world software development tasks. In addition, existing benchmarks often use poor code correctness metrics, providing misleading conclusions.
 
299
 
300
  * Check our [Leaderboard](https://lt-asset.github.io/REPOCOD/) for preliminary results using SOTA LLMs with RAG.
301
 
 
 
302
 
303
+ ```
304
  "instance_id": Instance ID in REPOCOD
305
  "version": Version of REPOCOD
306
  "gold_patches": {
 
322
  "code": Problem statement for code generation.
323
  "test": Problem statement for test generation.
324
  }
 
325
  "environment_setup_commit": base commit
326
  "evaluation": {
327
  "FAIL_TO_PASS": list of relevant test cases
328
  "PASS_TO_PASS": None, (all remaining tests that passes, we choose not to run the PASS_TO_PASS tests to avoid the computational cost)
329
  }
 
330
  ```
331
 
332
  ## Citation