nirajandhakal commited on
Commit
19e2192
·
verified ·
1 Parent(s): c8b4e9c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -28,6 +28,10 @@ StockZero learns to play chess by playing against itself. The core component is
28
 
29
  The model is trained using self-play data generated through MCTS, which guides the engine to explore promising game states.
30
 
 
 
 
 
31
  ### Input
32
 
33
  The model takes a chess board as input, represented as a 8x8x12 NumPy array. Each of the 12 channels in the input represent a specific piece type (Pawn, Knight, Bishop, Rook, Queen, King) for both white and black players, where each layer contains binary values.
 
28
 
29
  The model is trained using self-play data generated through MCTS, which guides the engine to explore promising game states.
30
 
31
+ This model card is for StockZero version 2 (v2) model. While the v1 model has same architecture, it had less self-play to learn policy. V1 model was played on only 20 self-play policy training for testing purposes to see whether the model will converge to lower value while v2 was played on 50 self-play games during policy training on Google Colaboratory Free Tier Notebook because larger self-play would result in high compute demand which is what I currently can't afford.
32
+
33
+ **Note**: StockZero v3 will be trained and open sourced soon.
34
+
35
  ### Input
36
 
37
  The model takes a chess board as input, represented as a 8x8x12 NumPy array. Each of the 12 channels in the input represent a specific piece type (Pawn, Knight, Bishop, Rook, Queen, King) for both white and black players, where each layer contains binary values.