NSW Police want access to Tinder’s sexual assault data. Cybersafety experts explain why it’s a date with disaster

Shutterstock
In summary
An analysis of NSW Police’s proposal to gain access to sexual assault reports on Tinder. Written for The Conversation by Kath Albury, Swinburne University of Technology, Rosalie Gillett, Queensland University of Technology and Zahra Zsuzsanna Stardust, Queensland University of Technology.
Dating apps have been under increased scrutiny for their role in facilitating harassment and abuse.
Last year an ABC investigation into Tinder found most users who reported sexual assault offences didn’t receive a response from the platform. Since then, the app has reportedly implemented new features to mitigate abuse and help users feel safe.
In a recent development, New South Wales Police announced they are in conversation with Tinder’s parent company Match Group (which also owns OKCupid, Plenty of Fish and Hinge) regarding a proposal to gain access to a portal of sexual assaults reported on Tinder. The police also suggested using artificial intelligence (AI) to scan users’ conversations for “red flags”.
Tinder already uses automation to monitor users’ instant messages to identify harassment and verify personal photographs. However, increasing surveillance and automated systems doesn’t necessarily make dating apps safer to use.
User safety on dating apps
Research has shown people have differing understandings of “safety” on apps. While many users prefer not to negotiate sexual consent on apps, some do. This can involve disclosure of sexual health (including HIV status) and explicit discussions about sexual tastes and preferences.
If the recent Grindr data breach is anything to go by, there are serious privacy risks whenever users’ sensitive information is collated and archived. As such, some may actually feel less safe if they find out police could be monitoring their chats.
Adding to that, automated features in dating apps (which are supposed to enable identity verification and matching) can actually put certain groups at risk. Trans and non-binary users may be misidentified by automated image and voice recognition systems which are trained to “see” or “hear” gender in binary terms.
Trans people may also be accused of deception if they don’t disclose their trans identity in their profile. And those who do disclose it risk being targeted by transphobic users.
Increasing police surveillance
There’s no evidence to suggest that granting police access to sexual assault reports will increase users’ safety on dating apps, or even help them feel safer. Research has demonstrated users often don’t report harassment and abuse to dating apps or law enforcement.
Consider NSW Police Commissioner Mick Fuller’s misguided “consent app” proposal last month; this is just one of many reasons sexual assault survivors may not want to contact police after an incident. And if police can access personal data, this may deter users from reporting sexual assault.

NSW Police Commissioner Mick Fuller got slammed by media and the public last month for suggesting a phone app could be used to record sexual consent. His suggestion came after reports of a growing number of sexual assault cases reported in the state. Dean Lewis/AAP
With high attrition rates, low conviction rates and the prospect of being retraumatised in court, the criminal legal system often fails to deliver justice to sexual assault survivors. Automated referrals to police will only further deny survivors their agency.
Moreover, the proposed partnership with law enforcement sits within a broader project of escalating police surveillance fuelled by platform-verification processes. Tech companies offer police forces a goldmine of data. The needs and experiences of users are rarely the focus of such partnerships.
Match Group and NSW Police have yet to release information about how such a partnership would work and how (or if) users would be notified. Data collected could potentially include usernames, gender, sexuality, identity documents, chat histories, geolocation and sexual health status.
The limits of AI
NSW Police also proposed using AI to scan users’ conversations and identify “red flags” that could indicate potential sexual offenders. This would build on Match Group’s current tools that detect sexual violence in users’ private chats.
While an AI-based system may detect overt abuse, everyday and “ordinary” abuse (which is common in digital dating contexts) may fail to trigger an automated system. Without context, it’s difficult for AI to detect behaviours and language that are harmful to users.
It may detect overt physical threats, but not seemingly innocuous behaviours which are only recognised as abusive by individual users. For instance, repetitive messaging may be welcomed by some, but experienced as harmful by others.
Also, even as automation becomes more sophisticated, users with malicious intent can develop ways to circumvent it.
If data are shared with police, there’s also the risk flawed data on “potential” offenders may be used to train other predictive policing tools.
We know from past research that automated hate-speech detection systems can harbour inherent racial and gender biases (and perpetuate them). At the same time we’ve seen examples of AI trained on prejudicial data making important decisions about people’s lives, such as by giving criminal risk assessment scores that negatively impact marginalised groups.
Dating apps must do a lot more to understand how their users think about safety and harm online. A potential partnership between Tinder and NSW Police takes for granted that the solution to sexual violence simply involves more law enforcement and technological surveillance.
And even so, tech initiatives must always sit alongside well-funded and comprehensive sex education, consent and relationship skill-building, and well-resourced crisis services.
This article is republished from The Conversation under a Creative Commons license. Read the original article.
-
Media Enquiries
Related articles
-
- Student News
- Law
Swinburne law students build crucial tech skills with AGL
Three Swinburne students have worked closely with AGL's legal team to develop a prototype workflow solution for internal legal requests.
Tuesday 15 March 2022 -
- Social Affairs
- Law
Are charities being silenced? Why a new law is alarming activists and could scuttle their election campaigns
Philanthropy expert Krystian Seibert explains how recent changes to electoral laws will impose more regulation on charities and other organisations that engage in the electoral process
Friday 03 December 2021 -
- Law
Australia’s housing laws are changing, but do they go far enough to prevent pet abandonment?
New South Wales has joined Queensland and the ACT in ending blanket bans on pets in apartments, but there is still too much uncertainty across the housing system. Analysis for The Conversation by Emma Power, Western Sydney University and Wendy Stone, Swinburne University of Technology.
Wednesday 15 September 2021 -
- Technology
- Law
- Business
Swinburne Law student develops legal tech solution for Spotify
Third-year student Nicola Jerkovic has partnered with Spotify as part of the Law Without Walls competition to create an innovative tech solution to a real problem facing the company’s legal team.
Friday 27 August 2021 -
- Law
- Politics
The government is clamping down on charities — and it could have a chilling effect on peaceful protest
The Australian government’s new regulations to crack down on “unlawful behaviour” may impact the work of legitimate charities. Analysis for The Conversation by Krystian Seibert, Swinburne University of Technology.
Friday 02 July 2021