But that's not what's actually happening. We've already been over this, either in this thread or another one, in which someone posted a pop-sci article claiming that AI researchers are baffled that ChatGPT "somehow learned biology even though we never taught it biology", and it turns out that if you ask an actual biologist to review ChatGPT's biology homework they'll tell you that no, it actually sucks at biology, and the same is true for any other of these "emergent abilities".
We are not only talking about ChatGPT 3, my comments were mainly about the AIs that are not wiped clean every time and can use feedback as its inputs.
Coding, specifically, is one field at which professional computer programmers have said repeatedly that the AI is fairly incompetent beyond some basic concepts.
That is quite different to what I've read.
If I recall, we practiced here by giving it beginner-level physics questions and watching it return completely wrong answers. The real takeaway is that "AI researchers" are probably not the best judge at how good an AI is at a subject or task that is outside their wheelhouse - as you would expect of any other kind of expert.
Not quite - it did indeed answer many questions correctly, including examination papers. But yes it also returned inaccurate answers.
I've managed to completely flummox ChatGPT
4 with a simple question "how many monarchs has the UK had since 1920?" It repeatedly brought back wrong information, whilst acknowledging it got it wrong. Eventually it simply stopped responding.... hmm.. the AI equivalent of "i
t will always politely apologize as programmed, and it will never one day decide "you know what, this guy is a rude bastard and I don't have to apologize if he's going to talk to me that way" ?
In the weeks after ChatGPT's release, when some people - mostly trolls trying as hard as they could to "break" the bot's imposed restrictions - started bringing attention to some embarrassingly bad or inappropriate responses they were able to get the bot to return, ChatGPT's engineers typically responded by patching the software to prevent those kinds of mistakes from recurring. They were usually able to deploy these corrective patches fairly quickly and effectively too, which doesn't seem like it should have been possible if the engineers legitimately had no idea how the bot generated responses.
From memory they did it in several ways - one was at the tokenisation stage - which is like the art AIs that won't accept words like "nude" as part of their input, get in at the input and you don't need to alter the black box, another way they also then added "monitoring" software to its
outputs. Earlier on you saw this in action when the app would start to answer a question and then literally erase what it had written and tell you "I'm sorry Dave, I'm afraid I can't do that." They did not re-run the learning process again which would be what was necessary to incorporate new rules.
This is why people keep finding ways around its restrictions, such jailbreaking is possible because it is not the "black box" that is being altered but the stuff created by people to block naughtiness and if you get around that you get access to the black-box.