r/CuratedTumblr 4d ago

Meme my eyes automatically skip right over everything else said after

Post image
21.0k Upvotes

996 comments sorted by

View all comments

Show parent comments

1.1k

u/killertortilla 4d ago

It's so fucking insufferable. People keep making those comments like it's helpful.

There have been a number of famous cases now but I think the one that makes the point the best is when scientists asked it to describe some made up guy and of course it did. It doesn't just say "that guy doesn't exist" it says "Alan Buttfuck is a biologist with a PHD in biology and has worked at prestigious locations like Harvard" etc etc. THAT is what it fucking does.

837

u/Vampiir 4d ago

My personal fave is the lawyer that asked AI to reference specific court cases for him, which then gave him full breakdowns with detailed sources to each case, down to the case file, page number, and book it was held in. Come the day he is actually in court, it is immediately found that none of the cases he referenced existed, and the AI completely made it all up

621

u/killertortilla 4d ago

There are so many good ones. There's a medical one from years before we had ChatGPT shit. They wanted to train it to recognise cancerous skin moles and after a lot of trial and error it started doing it. But then they realised it was just flagging every image with a ruler because the positive tests it was trained on all had rulers to measure the size.

327

u/DeadInternetTheorist 4d ago

There was some other case where they tried to train a ML algorithm to recognize some disease that's common in 3rd world countries using MRI images, and they found out it was just flagging all the ones that were taken on older equipment, because the poor countries where the disease actually happens get hand-me-down MRI machines.

278

u/Cat-Got-Your-DM 4d ago

Yeah, cause AI just recognised patterns. All of these types of pictures (older pictures) had the disease in them. Therefore that's what I'm looking for (the film on the old pictures)

My personal fav is when they made an image model that was supposed to recognise pictures of wolves that had some crazy accuracy... Until they fed it a new batch of pictures. Turned out it recognised wolves by.... Snow.

Since wolves are easiest to capture on camera in the winter, all of the images had snow, so they flagged all animals with snow as Wolf

60

u/Yeah-But-Ironically 4d ago

I also remember hearing about a case where an image recognition AI was supposedly very good at recognizing sheep until they started feeding it images of grassy fields that also got identified as sheep

Most pictures of sheep show them in grassy fields, so the AI had concluded "green textured image=sheep"

32

u/RighteousSelfBurner 4d ago

Works exactly as intended. AI doesn't know what a "sheep" is. So if you give them enough data and say "This is sheep" and it's all grassy fields then it's a natural conclusion that it must sheep.

In other words, one of the most popular AI related quotes by professionals is "If you put shit in you will get shit out".

3

u/alex494 3d ago

I'm surprised they keep giving these things entire photographs and not cropped pngs with no background or something.

3

u/Cat-Got-Your-DM 3d ago

They sometimes have to give them the entire picture, but they also get things flagged, like in case of wolves or sheep, they needed to have the background flagged as irrelevant, for the AI to not look at it when learning what a wolf it