nirajandhakal commited on
Commit
c8b4e9c
·
verified ·
1 Parent(s): e344866

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -1
README.md CHANGED
@@ -100,6 +100,17 @@ The model is intended for research, experimentation, and education purposes. Pot
100
  * The training data is limited to a small number of self-play games (50 games), therefore the strength of the engine is limited.
101
  * The model is trained on a single GPU, so longer training may require multi GPU support or longer runtime.
102
 
 
 
 
 
 
 
 
 
 
 
 
103
  ## How to Use
104
 
105
  ### Training
@@ -202,4 +213,5 @@ if np.sum(masked_policy_probs) > 0:
202
  masked_policy_probs /= np.sum(masked_policy_probs)
203
 
204
  print("Policy Output:", masked_policy_probs)
205
- print("Value Output:", value_output)
 
 
100
  * The training data is limited to a small number of self-play games (50 games), therefore the strength of the engine is limited.
101
  * The model is trained on a single GPU, so longer training may require multi GPU support or longer runtime.
102
 
103
+ ## Model Evaluation
104
+
105
+ This model was evaluated against a simple random move opponent using the `evaluate_model` method in the provided `evaluation_script.py`. The results are as follows:
106
+
107
+ * **Number of Games:** 200 (The model plays as both white and black in each game against the random agent.)
108
+ * **Win Rate:** 0.0150 (1.5%)
109
+ * **Draw Rate:** 0.6850 (68.5%)
110
+ * **Loss Rate:** 0.3000 (30%)
111
+
112
+ These scores indicate that the model, in its current state, is not a strong chess player. It draws a majority of games against a random opponent, but also loses a significant number. Further training and architecture improvements are needed to enhance its performance.
113
+
114
  ## How to Use
115
 
116
  ### Training
 
213
  masked_policy_probs /= np.sum(masked_policy_probs)
214
 
215
  print("Policy Output:", masked_policy_probs)
216
+ print("Value Output:", value_output)
217
+ ```