So as not to clog up other places, here is a thread for the general discussion of AI, its advances, implications, and suchlike. This thread meant primarily for the use of non-AIs (e.g. humans), but any AIs that happen by are of course welcome to join in. (If we're not sure whether someone is an AI or not, John Searle can presumably sort that out for us.)
As a starter, this article from today:
Do Large Language Models learn world models or just surface statistics?
It's a fun read, but the tl;dr is that the authors trained a GPT on transcripts of Othello games ("E5, D6, .." etc), so that it could successfully predict legal moves. Then they did various stuff with the GPT's internals, and were able to show that it had developed a "world model" of the game - i.e. that Othello has a grid of connected nodes whose state can alternate after each move - even though the GPT had only ever seen game transcripts.
Not a surprising result, exactly, but it's impressive that it can already be demonstrated, and it seems like this sort of poking around in AI models, and attempting to understand or intervene into their weights, will be a interesting thing to watch from here on out..