Releases: LeelaChessZero/lc0
Releases · LeelaChessZero/lc0
v0.29.0-rc1
In this release:
- New metal backend for apple systems. This is now the default backend for
macos builds. - New onnx-dml backend to use DirectML under windows, has better net
compatibility than dx12 and is faster than opencl. See the README for use
instructions, a separate download of the DirectML dll is required. - Full attention policy support in cuda, cudnn, metal, onnx, blas, dnnl, and
eigen backends. - Partial attention policy support in onednn backend (good enough for T79).
- Now the onnx backends can use fp16 when running with a network file (not with
.onnx model files). This is the default for onnx-cuda and onnx-dml, can be
switched on or off with by setting thefp16
backend option totrue
or
false
respectively. - The onednn package comes with a dnnl compiled to allow running on an intel gpu
by addinggpu=0
to the backend options. - The default net is now 791556 for most backends except opencl and dx12 that
get 753723 (as they lack attention policy support). - Support for using pgn book with long lines in training: selfplay can start at
a random point in the book. - New "simple" time manager.
- Support for double Fischer random chess (dfrc).
- Added TC-dependent output to the backendbench assistant.
- Starting with this version, the check backend compares policy for valid moves
after softmax. - Some assorted fixes and code cleanups.
v0.29.0-rc0
In this release:
- Initial support for attention policy, only cuda backend and partially in
blas/dnnl/eigen (good enough for T79). - Non multigather (legacy) search code and
--multigather
option are removed. - 15b default net is now 753723.
- The onnx backend now allows selecting gpu to use.
- Improved error messages for unsupported network files.
- Some assorted fixes.
v0.28.2
This is what should have been v0.28.1:
- Improved cuda performance for 512 filter networks on Amprere GPUs.
- Several fixes for the onnx backend.
- New lc0 modes to process network files: describenet, leela2onnx and onnx2leela
- Documentation updates.
- Correctness fixes for rescorer support functions.
v0.28.1-rc1
- Improved cuda performance for 512 filter networks on Amprere GPUs.
- Several fixes for the onnx backend.
- Command line options for network file conversion to/from onnx.
- Documentation updates.
- Correctness fixes for rescorer support functions.
v0.28.0
In this release:
- Multigather is now made the default (and also improved). Some search settings have changed meaning, so if you have modified values please discard them. Specifically,
max-collision-events
,max-collision-visits
andmax-out-of-order-evals-factor
have changed default values, but other options also affect the search. Similarly, check that your GUI is not caching the old values. - Updated several other default parameter values, including the MLH ones.
- Performance improvements for the cuda/cudnn backends. This includes the
multi_stream
cuda backend option that is off by default. You should test addingmulti_stream=true
tobackend-opts
(command line) or BackendOptions (UCI) if you have a recent GPU with a lot of VRAM. - Support for policy focus during training.
- Larger/stronger 15b default net for all packages except android, blas and dnnl that get a new 10b network.
- The distributed binaries come with the mimalloc memory allocator for better performance when a large tree has to be destroyed (e.g. after an unexpected move).
- The
legacy
time manager is again the default and will use more time for the first move after a long book line. - The
--preload
command line flag will initialize the backend and load the network during startup. This may help in cases where the GUI is confused by long start times, but only if backend and network are not changed via UCI options. - A 'fen' command was added as a UCI extension to print the current position.
- Experimental onednn backend for recent intel CPUs and GPUs.
- Added support for ONNX network files and runtime with the onnx backend.
- Several bug and stability fixes.
Note: Some small third-party nets seem to play really bad with the dx12 backend and certain GPU drivers, setting the enable-gemm-metacommand=false
backend option is reported to work around this issue.
v0.28.0-rc2
- The cuda backend option multi_stream is now off by default. You should
consider setting it to on if you have a recent gpu with a lot of vram. - Updated default parameters.
- Newer and stronger nets are included in the release packages.
- Added support for onnx network files and runtime with the "onnx" backend.
- Several bug and stability fixes.
v0.28.0-rc1
- Multigather is now made the default (and also improved). Some search settings
have changed meaning, so if you have modified values please discard them.
Specifically,max-collision-events
,max-collision-visits
and
max-out-of-order-evals-factor
have changed default values, but other options
also affect the search. Similarly, check that your gui is not caching the old
values. - Performance improvements for the cuda/cudnn backends.
- Support for policy focus during training.
- Larger/stronger 15b default net for all packages except android, blas and dnnl
that get a new 10b network. - The distributed binaries come with the mimalloc memory allocator for better
performance when a large tree has to be destroyed (e.g. after an unexpected
move). - The
legacy
time manager will use more time for the first move after a long
book line. - The
--preload
command line flag will initialize the backend and load the
network during startup. - A 'fen' command was added as a UCI extension to print the current position.
- Experimental onednn backend for recent intel cpus and gpus.
v0.27.0
v0.27.0-rc2
- Fix additional cases where 'invalid move' could be incorrectly reported.
- Replace WDL softmax in cudnn backend with same implementation as cuda
backend. This fixes some inaccuracy issues that were causing training
data to be rejected at a fairly low frequency. - Ensure that training data Q/D pairs form valid WDL targets even if there
is accumulated drift in calculation. - Fix for the calculation of the 'best q is proven' bit in training data.
- Multiple fixes for timelosses and infinite instamoving in smooth time
manager. Smooth time manager now made default after these fixes.
v0.27.0-rc1
- Fix a bug which meant
position ... moves ...
didn't work if the moves went off the end of the existing tree. (Which happens normally when playing from an opening book.)