← Back to context

Comment by dawnofdusk

8 months ago

Even if it's "implicit" I'm not sure if that matters that much. The point is that the model doesn't explicitly search anything, it just applies the learned transformation. If the weights of the learned transformation encode a sort of precomputed search and interpolation over the dataset, from an algorithmic perspective this still isn't search (it doesn't enumerate board states or state-action transitions).

>performing some function of the similarity of the input to those previously seen.

This is indeed what transformers do. But obviously it learns some sort of interpolation/extrapolation which lets it do well on board states/games outside the training set.

I agree, from a practical perspective what matters is that is distilling stockfish’s search to a certain extent, which could have computational efficiencies. “Without search” just means we’re not doing anything like minimax or MCTS.