Mr Blobby on Love Island, teapots with six-packs, Guy Fieri drowning in a vat of soup: AI-rendered images from the depths of our (cursed) imaginations have made us belly laugh ever since the tools came into our hands. Many of these images look janky, with blurring and odd textures that let us know thereâs no way they could ever be real.
Machine learning, however, is advancing faster than everâand programs like DALL-E 2, Stable Diffusion, and Midjourney are now able to quickly churn out photorealistic images in a matter of seconds. A series of striking polaroids featuring black female goth and heavy metal fans in the late 90s recently went viral on Twitter, showing groups of women in distinctive leather outfits, band tees, and rock hairstyles. Itâs hard to believe theyâre not actual pictures unearthed from sweaty gigs of the past, but are instead the work of a Midjourney user called Fallon Fox.
If you look closely, itâs quite easy to pinpoint the uncanny valley phenomenon: the handsâa tell-tale sign of machine learning-generated imagesâoften look wrong, sometimes with six-plus fingers gripping drinks weirdly. Some of the faces donât look quite right either, with out-of-place noses and teeth. But thereâs an overall ambience to the shots that feels very real. Nostalgic, even.
Other examples of this hyper-realistic photography being generated of the âpastâ are surfacing as more users play around with the technology. There are pictures from inner-city Glasgow in 1985 with trackie-clad lads standing in front of tower blocks, and 90s rival gangs from LA playing dice games in colour-coordinated fits. Another user produced a depiction of queer culture in the past. None of them look perfect, but when viewed as thumbnails or perceived without a second thought, your eyes could easily gloss over the flaws.
âIt is tricky [to differentiate human-shot photography from that created by AI], particularly when those photographs appear in contexts like social media, where our engagement with each image tends to be very brief,â Lewis Bush, a photographer and PhD researcher in Computational Photojournalism at the London School of Economics and Political Science, told SCREENSHOT. âThere are still clues in many of these generated images as to their origins, but they tend to be hard to pick up on without a concentrated look at the image.â
It seems like weâre going to be inundated with hyperreal AI-generated photography pretty soon too. Stock image company Shutterstock has just inked a deal to add images created by DALL-E 2. At the same time, in a bid to reassure aggrieved photographers, itâs offering reimbursements to those whose images the AI has been trained on.
If the internet gets saturated with these kinds of photographs, which sit alongside real historical documentation, are we in danger of creating false narratives about the past? Itâs worth noting that weâve been able to edit and doctor images for years, dating back to the pre-Photoshop retouching used by the Soviet Union, and image manipulation is nothing newâAI isnât doing anything that an advanced digital creator couldnât accomplish.
These AI images of Glasgow in 1985 are blowing my mind
— Michael M đ˝ (@michaelmphysics) November 22, 2022
đhttps://t.co/zE2fougfhD pic.twitter.com/57fj5RVZli
That being said, these tools are increasingly being used by (almost) the entire world rather than a comparatively tiny group of people. So, should we be worried about machine-generated imagery becoming more advanced? Jon McCormack, a generative art researcher at Monash University, Australia, doesnât think so. What he is concerned about however is âthe overall aesthetic and âsamenessâ of images generated using diffusion models.â
âBecause they are trained on images from the internet, they mimic popular styles and aesthetic values, diminishing representations that arenât as culturally homogeneous or ubiquitous,â McCormack shared. âSuch machine learning systems just reinforce this ubiquity and sameness, ultimately at the expense of cultural diversity and creativity.â
TikTok user and digital artist @eli_other, who uses DALL-E 2 to generate images using text prompts like âa LEGO set coming home to his wife cheating on himâ and âalbum cover for an all-horse metal band,â says heâs found that images generated by machines skew towards white male-centric themes. âFor many of my prompts, I use neutral language like âpersonâ and about 60 per cent of the time the âpersonâ is a white man. Alongside that, itâs about 20 per cent white women and 20 per cent people of colour of any gender.â
Reactions to the Black goth and heavy metal fans on Twitter ranged from âperfectâ and âbeautifulâ to âicky.â Non-white fans of that type of music did, of course, exist in the 90s, but the goth subculture has historically been associated with whiteness, and thereâs a question about whether misrepresentations of the past could lead to skewed narratives about historic oppression. Fallon Fox, the former MMA fighter whoâd used Midjourney to produce the polaroids, told Screen Rant that she was trying to âshow a representation of people like [herself],â a Black trans woman, in the metal scene. âI put a lot of references to 90s-era Black goths in there,â she added.
With most innovations comes some kind of pushback and, just like any other picture, Bush believes that itâs âimportant to understand what it is that you’re looking at, and how that image has been produced.â And there are endless benign and educational purposes for this kind of photo generation. One of Bushâs projects, titled An Antique Land zine, uses machine learning to picture London after its fall to the ravages of climate change with green foliage sprouting out of the Natural History Museum.