We used the 100 AI and 100 human White faces (half male, half female) from Nightingale and Farid. The AI faces were generated using StyleGAN2. The human faces were selected from the Flickr-Faces-HQ Dataset to match each of the AI faces as closely as possible (e.g., same gender, posture, and expression). All stimuli had blurred or mostly plain backgrounds, and AI faces were screened to ensure they had no obvious rendering artifacts (e.g., no extra faces in background). Screening for artifacts mimics how real-world users screen AI faces, either as scientists or for public use, and therefore captures the type and range of stimuli that appear online. Participants were asked to resize their screen so that stimuli had a visual angle of 12° wide × 12° high at ~50 cm viewing distance.
I don’t know why people (not saying you, more directed at the top commenter) keep acting like cherry picking AI images in these studies invalidate the results - cherry picking is how you use AI image generation tools, that’s why most will (or can) generate several at once so you can pick the best one. If a malicious actor was trying to fool people, of course they’d use the most “real” looking ones, instead of just the first to generate
Frankly the studies would be useless if they didn’t cherry pick, because it wouldn’t line up with real world usage
Tbh I’m more concerned about how they chose the human faces. I can’t explain it, but it feels like they were biased toward choosing ‘fake-looking’ faces, lol
The way it sounds right now is “AI generated faces don’t have all these artifacts 99% of the time” (I’m paraphrasing A LOT, but you get what I mean.)
The only way it sounds like that is if you don’t read the article at all and draw all your conclusions from just reading the title.
Don’t get me wrong, I’m sure many do just that, but that’s not the fault of the study. They clearly state their method for selecting (or “cherry picking”) images
deleted by creator
I don’t know why people (not saying you, more directed at the top commenter) keep acting like cherry picking AI images in these studies invalidate the results - cherry picking is how you use AI image generation tools, that’s why most will (or can) generate several at once so you can pick the best one. If a malicious actor was trying to fool people, of course they’d use the most “real” looking ones, instead of just the first to generate
Frankly the studies would be useless if they didn’t cherry pick, because it wouldn’t line up with real world usage
Tbh I’m more concerned about how they chose the human faces. I can’t explain it, but it feels like they were biased toward choosing ‘fake-looking’ faces, lol
deleted by creator
The only way it sounds like that is if you don’t read the article at all and draw all your conclusions from just reading the title.
Don’t get me wrong, I’m sure many do just that, but that’s not the fault of the study. They clearly state their method for selecting (or “cherry picking”) images
deleted by creator