Of course it is. That is the outcome of unsupervised learning.
It doesn’t have a sense of what’s true and false, what’s right and wrong.
It has learned to predict the next word very well and the prediction probability distribution was later tweaked with human feedback and automated test feedback.
They’ll train it to not predict these words as much (or basically at all) when this is asked. But a very large part of the model will not be touched by these interventions, so it will continue to predict text as it has learned to.
I'm really confused that people don't understand this. It's just predicting the most likely next text token and its trained on most internet text, so why would we expect anything at all different?
How do we know this is regurgitation and not something like an AI summary of top hits ala Bing chat? Is there a reference to source links, if not highly questionable
> It has learned to predict the next word very well and the prediction probability distribution was later tweaked with human feedback and automated test feedback.
It didn't predict anything here, it just ripped off a reddit comment.
It doesn’t have a sense of what’s true and false, what’s right and wrong.
It has learned to predict the next word very well and the prediction probability distribution was later tweaked with human feedback and automated test feedback.
They’ll train it to not predict these words as much (or basically at all) when this is asked. But a very large part of the model will not be touched by these interventions, so it will continue to predict text as it has learned to.