After the quoted thread generated furious debate about whether Bing AI keeps track of board position when given a chess game, it occurred to me to just ask it! It got 63 of 64 squares correct; it missed the pawn on b5. Accuracy degrades as the game gets longer.
Wouldn't the fact that accuracy degrades suggest that it is not an abstracted model or internal representation but simply a probabilistic reply to your very specific prompt?
Reasonable people can disagree and I know I won't change anyone's mind, but here's why I see it differently. Gradual accuracy degradation suggests an internal representation exists but with a small chance of error in updating it, and that error accumulates with move count. 1/n
Without an internal representation, I think accuracy would drop exponentially, and there's no way it would get to 25 ply without losing the plot completely. I don't know any probabilistic mechanism that can simulate state updates without actually doing state updates.
I suspect degradation is just due to the context size limitation. You should try it again in a couple of weeks.
Maybe you are correct, but the inner monologue can be much longer than the output - and I suspect it is in this case, though I haven't looked.
Mar 4, 2023 · 6:36 PM UTC