Hier der Link zur Training Seite:
https://training.lczero.org/So werden auf Discord die Engines und Netze veröffentlicht:
Big transformer 3. New network arch using smolgen-augmented self-attention from BT2. It has embedding size 768, ffn projection size 1024, 24 heads per layer, and 15 smolgen encoder layers with mish activation. There are also cuda optimizations available, which should reduce latency by 10 to 15%. It has 3 policy heads: vanilla, optimistic and soft. Vanilla and optimistic can be used for play, while soft helps speedup training. The optimistic policy head improves policy predictions drastically in tactical positions. It has 3 value heads: winner, q and st. Winner head is trained on game outcome, while q head is trained on position Q-value. The ST value head is a weighted average of short-term future value from current position. The ideas are from Katago methods:
https://github.com/lightvector/KataGo/blob/master/docs/KataGoMethods.md#optimistic-policy.Quickstart:
The new engine version can be found at
https://github.com/Ergodice/lc0/tree/masterExecutable files can be found at
https://ci.appveyor.com/project/Ergodice/lc0/builds/48346860BT3 can be found at
https://storage.lczero.org/files/networks-contrib/BT3-768x15x24h-swa-2790000.pb.gzTo enable all the new features, put the following in the config file:
```
--uncertainty-weighting-cap=1.03
--uncertainty-weighting-coefficient=0.13
--uncertainty-weighting-exponent=-1.76
--use-uncertainty-weighting=true
```
If you are using a single GPU, add
--backend-opts=policy_head=vanilla,value_head=winner
Otherwise, check the Github for instructions.
Yaml specification at
https://discord.com/channels/425419482568196106/1192835078099828895Last updated: 3/7/2024