Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's a knowledge distillation. You can then use this smaller, more efficient models instead of the larger one.


Or maybe it is just memorizing a very large number of games.


They address the possibility of memorization in the PDF:

> This effect cannot be explained by memorization since < 1.41% of the initial puzzle board states appear in our training set.


Seems more like a 'compression' of the large number of games, or even like an approximate 'index' of the database


Is this network smaller than stockfish and by what metric is that?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: