Deep Dives Level Up Newsletters Saved Articles Challenges

From promoting anti-trans ideologies to severing ties with LGBT+ groups, is the EHRC still fit for its role?

By Louis Shankar

Feb 22, 2022

COPY URL

The Equality and Human Rights Commission (EHRC) is a non-departmental public body in England and Wales that was established by the Equality Act in 2006. Meanwhile, Scotland and Northern Ireland have separate bodies working towards the same end. Lately, however, the EHRC has been under fire for promoting views of anti-trans lobby groups while ignoring trans and LGBT+ rights groups. The Wikipedia page for the Commission currently reads: “Since 2021, the organisation has increasingly targeted and worked against LGBT rights, especially transgender rights.”

Last year, a former Chair of the EHRC said that the appointments made by Liz TrussMinister for Women and Equalities since 2019, as well as now Foreign Secretary—weakened its independence. Liberty, the UK’s largest civil liberties organisation, has said that the EHRC’s status as a national human rights institution should be “reviewed.”

Baroness Kishwer Falkner, the EHRC Chair, recently wrote to the Scottish government to advise further consideration before undertaking reform of gender recognition in Scotland—despite not one but two large-scale public consultations—as well as this law being part of devolved powers and, thus, beyond her remit. EHRC Scotland came out in favour of significant reform several years ago, following the initial public consultation.

19 of the UK’s major LGBT+ rights groups have, in recent months, formally severed ties with the EHRC. These groups are now in a coalition led by Stonewall appealing to the Global Alliance of National Human Rights Institutions (GANHRI) calling for the body to lose its ‘A-rating’—which would be an unprecedented move. In an open letter, dozens of leaders of trans and LGBTQ+ organisations wrote: “The EHRC has a wide range of powers it could use to make equality a reality for our communities, whether influencing public policy debates, strategically litigating to advance our rights, catalysing change in complex policy areas like healthcare or providing guidance that resolves uncertainty and ambiguity about delivering our rights in reality.”

They continued: “It was a kick in the teeth to trans people to see the EHRC appear to put their organisational weight behind a movement that has only contributed to rising hate for trans people in communities, creating a policy environment where it is harder for trans people to access their rights.”

Ben Hunte—formerly at the BBC but now reporting for VICE—has broken several stories about suspicious and prejudiced behaviour at the EHRC. Although they claim to reach out to groups representing a variety of positions, it seems they have given preferential treatment to vocally anti-trans groups, such as the LGB Alliance and Fair Play For Women.

Recent emails, some sent from Falkner’s parliamentary address and not the official EHRC one, show collusion with various anti-trans organisations. For example, Falkner deferred a response to questions from the Sunday Times in order to wait and hear the opinions of Fair Play To Women. VICE has also reported that staff are quitting the EHRC due to institutionalised transphobia, with current and former staff members describing an “anti-LGBT” culture at the commission.

This past week, The Guardian published an op-ed by Falkner, in which she defended her position—building on recent official statements published via the EHRC’s Twitter account that sought to “downplay, minimise and deny the recent revelations.” In the piece, Falkner provided no evidence to support her rebuttals and repeated multiple meaningless platitudes. She made no reference to the fact that the UK’s major LGBT+ and trans rights groups have severed ties with her organisation. There have been many and repeated calls for Falkner to resign—although her replacement would, similarly, be installed by Truss.

Another appointee currently on the board of the EHRC is Alasdair Henderson, a barrister who took part in a landmark case regarding trans rights in the UK. In 2020, he represented Keira Bell, a young woman who took legal action against the NHS to halt the use of puberty blockers in children aged 16 and under. The case was, initially, successful and the NHS removed all medical treatment for trans children—until the decision was overturned by appeal nearly a year later. On Twitter, Henderson has also liked and reposted tweets criticising Black Lives Matter protesters and has described words such as ‘misogynist’ and ‘homophobe’ as “highly ideological propaganda terms.”

All this comes at a moment when LGBT+ rights are under threat across Europe and beyond. ILGA-Europe revealed many serious concerns for LGBT+ rights in the UK in their 2022 Annual Report, noting: “Anti-trans rhetoric continued to cause serious damage in the UK again this year. Civil society reports that mainstream newspapers ran one or more anti-trans articles every day.” The Council of Europe recently approved a motion that included the acknowledgement of “virulent attacks on the rights of LGBTI people that have been occurring for several years in, amongst other countries, Hungary, Poland, the Russian Federation, Turkey and the United Kingdom.” The fact that the UK is placed among such company, where anti-LGBT+ legislation and violence is rife, should be of concern to us all.

The LGBT+ community of the UK needs support to fight for our rights and the EHRC—in its current state, with its current leadership—is fundamentally unsuited to this role.

EU prompted to ban automated recognition of gender and sexual orientation for putting LGBTQ+ lives at risk

By Yair Oded

Apr 30, 2021

COPY URL

“I have enough problems going around the world [as a trans person] without literal buildings constantly telling me, ‘Hey, hey, I think you’re a dude’,” Os Keyes, a gender and technology researcher based at the University of Washington told Screen Shot. Keyes was referring to the growing trend of governments and companies deploying automated recognition of gender and sexual orientation in order to identify citizens and consumers in a wide variety of spaces, from airport terminals, retail stores and billboards to social media platforms and mobile applications.

This software, which attempts to classify people as either ‘male’ or ‘female’ based on their facial features, the way they sound and the manner in which they move, places those whose gender doesn’t match the sex they were assigned at birth at great risk of further marginalisation, exclusion and discrimination. Harnessing the rising ubiquity of AI systems, automated gender recognition technology also threatens to reinforce outdated social taboos and stereotypes surrounding gender and effectively erase anything existing outside of the crudest binary perception of ‘male’ and ‘female’.

As the EU embarks on a legislative process of regulating the use of AI within the Union, a joint campaign launched by All Out, Access Now, Reclaim Your Face and Os Keyes is calling on the EU to include an explicit ban on automated gender and sexual orientation recognition in the bill.

On 21 April, the EU Commission—the executive branch of the EU—delivered its proposal for a legal framework to regulate AI. While it did highlight the inherent risks of some AI applications, the Commission did not go as far as prohibiting the deployment of automated gender recognition. The joint campaign to ban the technology, which so far has gained over 24,000 signatures, will now place its focus on the EU Parliament and Council, which are slated to continue working on the AI regulation bill.

The campaign originally stemmed from Keyes’ research about gender recognition systems and their impact on trans and nonbinary people. “I was prompted to study these gender recognition algorithms by having to see them used in my own discipline […] seeing people use it for research purposes and as a consequence producing research that cut out people who these systems cannot recognise,” Keyes told Screen Shot. “As I got in further,” they added, “I got to see more examples of it being used and deployed in the real world and a lot of people talking about deploying it further in situations that seem very, very dangerous for trans and gender non-conforming people.”

Keyes’ research was then referenced in the EU’s five-year LGBTI strategy, in a passage pointing out the danger in deploying automated gender recognition.

When Yuri Guaiana, senior campaign manager at All Out—an international LGBTQI advocacy organisation—came across Keyes’ quote in the EU’s LGBTI strategy he became fascinated with the topic and upon further research had launched a campaign to pressure the EU to ban automated gender and sexual orientation recognition. To that end, All Out joined forces with Access Now, an NGO advocating for a human rights-based regulation of AI, and Reclaim Your Face, a citizen initiative to ban biometric mass surveillance in the EU. They also got the endorsement of Keyes, who signed the letter submitted to the EU Commission along with the petition.

Speaking to Screen Shot, Keyes mentioned various existing applications of automated gender and sexual orientation recognition and highlighted some of the risks this technology poses for trans and gender non-conforming people.

One of the examples they referenced was a campaign by the Berlin Metro on International Women’s Day 2019, where women could pay 21 per cent less than men for a ticket. In order to authenticate a rider’s gender, automated gender recognition software was embedded in ticketing machines; those who failed to be recognised as female by the system were instructed to seek help from a service person at the station.

Keyes has pointed out two main issues in this case: “the first is the fact that you are being told ‘no you do not fit’,” they said. “The second is this idea of ‘well you can just go talk to an employee and they’ll work it out for you’,” they added. “Queer and trans people do not have the best experiences going to officials going ‘hey, just to let you know, I don’t fit, and I’m not meant to be here, and can you please fix this’. And when we think about the proposed deployments in places like bathrooms, you can see pretty clearly how that could get a lot more harmful and difficult.”

Keyes also mentioned the growing use of this technology in advertising, including on physical billboards that curate ads based on the perceived gender of the person walking past it: cars for men, dresses for women, and so on. Keyes pointed out that beyond the harm this application of automated gender recognition could cause trans and non-binary people, it also circulates incredibly negative and limiting social messages pertaining to gender: “This is what you’re allowed to do with gender, this is who you can be, this is what you can buy,” they said. Yuri Guaiana of All Out seconds this analysis. “How are you assuming that just because of your gender you are interested in certain products?” he said, highlighting that “interests are more important than gender in consumer behaviour.”

But Keyes emphasised the particular trauma this type of advertising can inflict on trans and gender non-conforming people. To them, the high potential of such advertising tools to misgender people who do not ‘fall neatly’ into either gender category and its implied message that they simply do not fit embody a blatant manifestation of transphobia. “What [transphobia] actually looks like is lots of small interactions […] it’s a death of a thousand cuts.” Keyes said. “And this is something I think anyone who is trans experiences on a day-to-day basis, like the constant small harms.”

Another application of the technology, which Keyes maintains is rarer but certainly existent, is in passport biometrics and various authentication systems. In this type of deployment, automated gender recognition is used to try and reduce the number of face images the given machine has to sort through in order to confirm the person’s identity. “The problem with this is if it gets it wrong, one way or the other, then what you get is the system concluding that this person does not appear in the database even though they do, and […] someone [could be] locked out of the system for being gender non-conforming,” Keyes said, adding that the secrecy with which this technology is shrouded and the lack of transparency regarding where, when and how it is being deployed amplifies its risk.

“We know that everyone is talking about doing it, and they most certainly are, but we can’t tell where and we can’t tell which discriminatory outcomes are caused by this,” they said, referencing a case where a trans woman’s identity could not be verified by Uber’s algorithm. “That could look a hell of a lot worse if we were talking about places like, again, biometrics, border control, passport security systems; places where you have much fewer rights or abilities to appeal if you can’t even work out what the system is not recognising about you in the first place […] and where the consequences of forced interactions with officials can be much more strenuous.”

Delineating the broader harm automated gender and sexual orientation recognition can inflict, Guaiana of All Out mentioned that the use of this technology could prove life-threatening in countries where being LGBTQI is illegal. “If they are using [automated gender and sexual orientation recognition] in places where being gay is illegal, and they can predict with a huge margin of being wrong that somebody rallying against something or walking in the street is gay—that can have very serious consequences,” Guaiana said. “This technology is used by government agents, but also private companies. It is censorship. Because in certain countries […] they could start surveilling people just because they predicted they are LGBTI.”

After reading over the EU Commission’s proposal last week, Guaiana, as well as other members of the campaign, noted that despite listing some applications of AI that should be prohibited, the Commission did not go as far as it should have in calling for a ban on harmful AI technologies that violate fundamental rights. “There is no explicit—or implicit, for that matter—ban on automatic recognition of gender and sexual orientation. For us, of course, this needs improvement,” Guaiana told Screen Shot. 

But All Out and its partners are far from discouraged. “Of course we would have preferred very much for the Commission to put [the ban] in the initial draft,” said Guaiana, “but I think it’s going to be a lengthy legislative process, [and] it’s still a good starting point […] There is still room to grow the campaign, keep the pressure up, and finally win this battle.”

Once more signatures are gathered and the legislative agenda and timeline of the EU Parliament and Council become known, the campaign to ban automated recognition of gender and sexual orientation will direct its resources at the Union’s representatives, recognising that they have the authority to amend the Commission’s recommendation and introduce the ban into the bill.

Guaiana and the other organisers of the campaign all believe that a ban on this particular type of technology in the EU could possibly have a global ripple effect, as did the General Data Protection Regulation (GDPR) back in 2016. Such a prohibition, says Guaiana, could “Help forbid the EU not only from implementing this technology within the EU, but also from exporting it […] and therefore that can help slow down the spread of this technology around the world.”

As we tackle the behemoth that is the tech industry, and as we try to regulate the application of various AI technologies and their deployment by both governments and companies, it is easy to feel powerless in the face of their seemingly inexorable force. Keyes, however, offers a slightly more optimistic—though pragmatist, as they define it—take on the issue. “I happen to believe that people thinking they can’t interfere [with technological development] is why interfering hasn’t worked thus far,” they said, “and there are a lot of examples that we don’t necessarily think about of technologies being banned in ways that did seriously derail things. Like, I’m a trans person, do you know how shitty trans healthcare is partly because nobody bothered doing any research because of the social taboos behind it?”

“We think of them as bad examples, but in a weird way they actually demonstrate that we can intervene in technological development; we can slow things down and we can redirect things,” they said, adding that our objective shouldn’t only be to root out the already existing technologies that prove harmful, but challenge the very way we approach, research and develop technology in the first place. “I think it’s possible,” they finally said, “because, well, if changing how people do things isn’t possible then the technology industry isn’t shit, because that’s what they claim they’ve been doing this whole time. Like, you’re telling me that your app can disrupt society beyond recognition, but also your software developers’ workflow is immutable and cannot be changed? One of those two things is false.”

 

×

Emails suck! Ours don't

Sign up to our weekly newsletter

 

Don't show again