Umeria wrote:Nilokeras wrote:If I know for a fact that there are 5329 sparrows in 2019 and I run the model and it says in 2019, based on environmental conditions, reproductive rates/mortality rates, food availability, human impacts and whatever else, there are 10 million of them, the model is simply wrong, and we go to understand why it isn't working. When ChatGPT does it, AI researchers insist it's 'hallucinating', and clasp their hands to their cheeks and go 'now isn't that interesting'. It's not, it's a model doing what models do - be wrong. And once you understand that or don't try to delude yourself into thinking its special, it's hard not to see all of this as self-serving pablum.
Ah, they're ignoring one of the basic principles of programming: "If there's a bug, it's because you put it there."
It's not even a bug, per se. Statistical models, by their nature, are not always accurate - we try to minimize the rate in which the model predicts the wrong value or outcome, but in some small percentage of times it runs there will be an incorrect prediction. In ecology we try to be very fastidious about keeping track of how often this happens, report accurately on the rate and design the models to minimize them.
ChatGPT is no different here: sometimes because it is using a very complicated statistical technique to predict which words should be assembled in which order, it gets it wrong. It'll say 'logistical regression' instead of 'beta regression'. Which is a problem in and of itself when people use it as a general-purpose search engine or Quora replacement, because it will state the wrong answer to a question very authoritatively.
The other type of 'error' of course is when people purposefully ask it an odd or leading question, and it confidently creates a strange pastiche of an answer - like this from the WIki page:
It is, after all, building its response based on the words you provide it, so if 'churro' , 'scientists' and 'surgery' in a prompt, it will obviously incorporate them into its response, which is a mash-up of a scientific press release or news article and descriptions of food or recipes. It's not a search engine, after all: it doesn't incorporate info about context, source or even the ultimate veracity of its responses. It is purely and entirely looking at the statistical relationships between the words you give it and the words in its corpus.
If this were an academic curiosity it might be interesting by itself, again, but there's a whole entire billion dollar tech industry looking for The Next Big thing and a coterie of parasites, rubes and hucksters have latched onto it, treating it as though it were some sort of alien life form or advanced AI when it gives them spooky answers to a spooky prompt. In reality it's an iterative and not particularly original product of the infinite money and computational power Silicon Valley can throw at a problem to solve it clumsily.