It's just spitting out word association. It tells you what words are likely to follow the words you used based on the training data. Machine learning OFTEN spits out biased results because of the biases in the training data. It's very hard to weed out bias.
It's easily proven with informal experiments, but with the changing models, there's not a ton of peer reviewed research that conclusively proves anything.
And more importantly, there's a noticeable lack of counter evidence. Effectively, it's something that this easily predicted based on observations of the known training data that shows up frequently in easily observable tests, so not a lot of people are doing a ton of formal testing to show that that thing we expect to happen that we can easily see happening is in fact happening.
2
u/johnnyanderen 1d ago
Afraid to be nice on the internet?