krishnateja95 commited on
Commit
358ba4a
·
verified ·
1 Parent(s): a229d2e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -49
README.md CHANGED
@@ -173,9 +173,7 @@ The model was evaluated on the OpenLLMv1 leaderboard task, using [lm-evaluation-
173
 
174
 
175
 
176
-
177
  ### Accuracy
178
-
179
  <table>
180
  <thead>
181
  <tr>
@@ -191,97 +189,97 @@ The model was evaluated on the OpenLLMv1 leaderboard task, using [lm-evaluation-
191
  <tr>
192
  <td rowspan="7"><b>OpenLLM V1</b></td>
193
  <td>ARC-Challenge (Acc-Norm, 25-shot)</td>
194
- <td>61.01</td>
195
- <td>60.58</td>
196
- <td>xyz</td>
197
  </tr>
198
  <tr>
199
  <td>GSM8K (Strict-Match, 5-shot)</td>
200
- <td>70.51</td>
201
- <td>70.96</td>
202
- <td>xyz</td>
203
  </tr>
204
  <tr>
205
  <td>HellaSwag (Acc-Norm, 10-shot)</td>
206
- <td>80.74</td>
207
- <td>80.35</td>
208
- <td>xyz</td>
209
  </tr>
210
  <tr>
211
  <td>MMLU (Acc, 5-shot)</td>
212
- <td>abc</td>
213
- <td>ijk</td>
214
- <td>xyz</td>
215
  </tr>
216
  <tr>
217
  <td>TruthfulQA (MC2, 0-shot)</td>
218
- <td>54.55</td>
219
- <td>54.02</td>
220
- <td>xyz</td>
221
  </tr>
222
  <tr>
223
  <td>Winogrande (Acc, 5-shot)</td>
224
- <td>78.37</td>
225
- <td>77.51</td>
226
- <td>xyz</td>
227
  </tr>
228
  <tr>
229
  <td><b>Average Score</b></td>
230
- <td><b>abc</b></td>
231
- <td><b>ijk</b></td>
232
- <td><b>xyz</b></td>
233
  </tr>
234
  <!-- OpenLLM Leaderboard V2 -->
235
  <tr>
236
  <td rowspan="7"><b>OpenLLM V2</b></td>
237
  <td>IFEval (Inst Level Strict Acc, 0-shot)</td>
238
- <td>abc</td>
239
- <td>60.19</td>
240
- <td>xyz</td>
241
  </tr>
242
  <tr>
243
  <td>BBH (Acc-Norm, 3-shot)</td>
244
- <td>abc</td>
245
- <td>ijk</td>
246
- <td>xyz</td>
247
  </tr>
248
  <tr>
249
  <td>Math-Hard (Exact-Match, 4-shot)</td>
250
- <td>abc</td>
251
- <td>ijk</td>
252
- <td>xyz</td>
253
  </tr>
254
  <tr>
255
  <td>GPQA (Acc-Norm, 0-shot)</td>
256
- <td>abc</td>
257
- <td>32.55</td>
258
- <td>xyz</td>
259
  </tr>
260
  <tr>
261
  <td>MUSR (Acc-Norm, 0-shot)</td>
262
- <td>abc</td>
263
- <td>39.81</td>
264
- <td>xyz</td>
265
  </tr>
266
  <tr>
267
  <td>MMLU-Pro (Acc, 5-shot)</td>
268
- <td>abc</td>
269
- <td>37.50</td>
270
- <td>xyz</td>
271
  </tr>
272
  <tr>
273
  <td><b>Average Score</b></td>
274
- <td><b>abc</b></td>
275
- <td><b>ijk</b></td>
276
- <td><b>xyz</b></td>
277
  </tr>
278
- <!-- HumanEval -->
279
  <tr>
280
- <td rowspan="2"><b>Coding</b></td>
281
  <td>HumanEval Pass@1</td>
282
- <td>abc</td>
283
- <td>ijk</td>
284
- <td><b>xyz</b></td>
285
  </tr>
286
  </tbody>
287
  </table>
 
173
 
174
 
175
 
 
176
  ### Accuracy
 
177
  <table>
178
  <thead>
179
  <tr>
 
189
  <tr>
190
  <td rowspan="7"><b>OpenLLM V1</b></td>
191
  <td>ARC-Challenge (Acc-Norm, 25-shot)</td>
192
+ <td>60.92</td>
193
+ <td>60.92</td>
194
+ <td>100.00</td>
195
  </tr>
196
  <tr>
197
  <td>GSM8K (Strict-Match, 5-shot)</td>
198
+ <td>71.11</td>
199
+ <td>70.66</td>
200
+ <td>99.36</td>
201
  </tr>
202
  <tr>
203
  <td>HellaSwag (Acc-Norm, 10-shot)</td>
204
+ <td>80.75</td>
205
+ <td>80.48</td>
206
+ <td>99.67</td>
207
  </tr>
208
  <tr>
209
  <td>MMLU (Acc, 5-shot)</td>
210
+ <td>68.20</td>
211
+ <td>67.96</td>
212
+ <td>99.64</td>
213
  </tr>
214
  <tr>
215
  <td>TruthfulQA (MC2, 0-shot)</td>
216
+ <td>54.54</td>
217
+ <td>54.18</td>
218
+ <td>99.34</td>
219
  </tr>
220
  <tr>
221
  <td>Winogrande (Acc, 5-shot)</td>
222
+ <td>78.45</td>
223
+ <td>78.14</td>
224
+ <td>99.60</td>
225
  </tr>
226
  <tr>
227
  <td><b>Average Score</b></td>
228
+ <td><b>69.00</b></td>
229
+ <td><b>68.72</b></td>
230
+ <td><b>99.59</b></td>
231
  </tr>
232
  <!-- OpenLLM Leaderboard V2 -->
233
  <tr>
234
  <td rowspan="7"><b>OpenLLM V2</b></td>
235
  <td>IFEval (Inst Level Strict Acc, 0-shot)</td>
236
+ <td>81.89</td>
237
+ <td>81.41</td>
238
+ <td>99.41</td>
239
  </tr>
240
  <tr>
241
  <td>BBH (Acc-Norm, 3-shot)</td>
242
+ <td>50.70</td>
243
+ <td>50.96</td>
244
+ <td>100.51</td>
245
  </tr>
246
  <tr>
247
  <td>Math-Hard (Exact-Match, 4-shot)</td>
248
+ <td>20.24</td>
249
+ <td>20.77</td>
250
+ <td>102.61</td>
251
  </tr>
252
  <tr>
253
  <td>GPQA (Acc-Norm, 0-shot)</td>
254
+ <td>29.53</td>
255
+ <td>29.95</td>
256
+ <td>101.42</td>
257
  </tr>
258
  <tr>
259
  <td>MUSR (Acc-Norm, 0-shot)</td>
260
+ <td>38.89</td>
261
+ <td>38.62</td>
262
+ <td>99.32</td>
263
  </tr>
264
  <tr>
265
  <td>MMLU-Pro (Acc, 5-shot)</td>
266
+ <td>37.71</td>
267
+ <td>37.48</td>
268
+ <td>99.38</td>
269
  </tr>
270
  <tr>
271
  <td><b>Average Score</b></td>
272
+ <td><b>43.16</b></td>
273
+ <td><b>43.20</b></td>
274
+ <td><b>100.09</b></td>
275
  </tr>
276
+ <!-- Coding -->
277
  <tr>
278
+ <td rowspan="1"><b>Coding</b></td>
279
  <td>HumanEval Pass@1</td>
280
+ <td>N/A</td>
281
+ <td>N/A</td>
282
+ <td>N/A</td>
283
  </tr>
284
  </tbody>
285
  </table>