Update README.md
Browse files
README.md
CHANGED
@@ -100,6 +100,17 @@ The model is intended for research, experimentation, and education purposes. Pot
|
|
100 |
* The training data is limited to a small number of self-play games (50 games), therefore the strength of the engine is limited.
|
101 |
* The model is trained on a single GPU, so longer training may require multi GPU support or longer runtime.
|
102 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
103 |
## How to Use
|
104 |
|
105 |
### Training
|
@@ -202,4 +213,5 @@ if np.sum(masked_policy_probs) > 0:
|
|
202 |
masked_policy_probs /= np.sum(masked_policy_probs)
|
203 |
|
204 |
print("Policy Output:", masked_policy_probs)
|
205 |
-
print("Value Output:", value_output)
|
|
|
|
100 |
* The training data is limited to a small number of self-play games (50 games), therefore the strength of the engine is limited.
|
101 |
* The model is trained on a single GPU, so longer training may require multi GPU support or longer runtime.
|
102 |
|
103 |
+
## Model Evaluation
|
104 |
+
|
105 |
+
This model was evaluated against a simple random move opponent using the `evaluate_model` method in the provided `evaluation_script.py`. The results are as follows:
|
106 |
+
|
107 |
+
* **Number of Games:** 200 (The model plays as both white and black in each game against the random agent.)
|
108 |
+
* **Win Rate:** 0.0150 (1.5%)
|
109 |
+
* **Draw Rate:** 0.6850 (68.5%)
|
110 |
+
* **Loss Rate:** 0.3000 (30%)
|
111 |
+
|
112 |
+
These scores indicate that the model, in its current state, is not a strong chess player. It draws a majority of games against a random opponent, but also loses a significant number. Further training and architecture improvements are needed to enhance its performance.
|
113 |
+
|
114 |
## How to Use
|
115 |
|
116 |
### Training
|
|
|
213 |
masked_policy_probs /= np.sum(masked_policy_probs)
|
214 |
|
215 |
print("Policy Output:", masked_policy_probs)
|
216 |
+
print("Value Output:", value_output)
|
217 |
+
```
|