THIS ARTICLE/PRESS RELEASE IS PAID FOR AND PRESENTED BY NTNU Norwegian University of Science and Technology - read more
Algorithms can prevent online abuse
The number of abuse cases against children via the internet has increased by almost 50 per cent in five years, according to the Norwegian Broadcasting Corporation. Researchers at NTNU in Gjøvik have developed algorithms that can help detect planned online grooming by analysing conversations.
Millions of children log into chat rooms every day to talk with other children. One of these 'children' could well be a man pretending to be a 12-year-old girl with far more sinister intentions than having a chat about My Little Pony episodes.
Inventor and NTNU professor Patrick Bours at AiBA is working to prevent just this type of predatory behaviour. AiBA, an AI-digital moderator that Bours helped found, can offer a tool based on behavioural biometrics and algorithms that detect sexual abusers in online chats with children.
And now, as recently reported by Dagens Næringsliv, a national financial newspaper, the company has raised capital of NOK 7.5. million, with investors including Firda and Wiski Capital, two Norwegian-based firms.
In its newest efforts, the company is working with 50 million chat lines to develop a tool that will find high-risk conversations where abusers try to come into contact with children. The goal is to identify distinctive features in what abusers leave behind on gaming platforms and in social media.
“We are targeting the major game producers and hope to get a few hundred games on the platform,” Hege Tokerud, co-founder and general manager, told Dagens Næringsliv.
Cyber grooming a growing problem
Cyber grooming is when adults befriend children online, often using a fake profile.
However, “some sexual predators just come right out and ask if the child is interested in chatting with an older person, so there’s no need for a fake identity,” Bours says.
The perpetrator’s purpose is often to lure the children onto a private channel so that the children can send pictures of themselves, with and without clothes, and perhaps eventually arrange to meet the young person.
The perpetrators don’t care as much about sending pictures of themselves, Bours explains.
“Exhibitionism is only a small part of their motivation. Getting pictures is far more interesting for them, and not just still pictures, but live pictures via a webcam," he says. "Overseeing all these conversations to prevent abuse from happening is impossible for moderators who monitor the system manually. What’s needed is automation that notifies moderators of ongoing conversation."
AiBA has developed a system using several algorithms that offers large chat companies a tool that can discern whether adults or children are chatting. This is where behavioural biometrics come in.
An adult male can pretend to be a 14-year-old boy online. But the way he writes – such as his typing rhythm, or his choice of words – can reveal that he is an adult man.
Machine learning key
The AiBA tool uses machine learning methods to analyse all the chats and assess the risk based on certain criteria. The risk level might go up and down a little during the conversation as the system assesses each message. The red warning symbol lights up the chat if the risk level gets too high, notifying the moderator who can then look at the conversation and assess it further.
In this way, the algorithms can detect conversations that should be checked while they are underway, rather than afterwards when the damage or abuse might have already occurred. The algorithms thus serve as a warning signal.
Cold and cynical
Bours analysed lots of chat conversations from old logs to develop the algorithm.
“By analysing these conversations, we learn how such men ‘groom’ the recipients with compliments, gifts and other flattery, so that they reveal more and more. It’s cold, cynical and carefully planned,” he says. “Reviewing chats is also a part of the learning process such that we can improve the AI and make it react better in the future.”
“The danger of this kind of contact ending in an assault is high, especially if the abuser sends the recipient over to other platforms with video, for example. In a live situation, the algorithm would mark this chat as one that needs to be monitored," he says.
Analysis in real time
“The aim is to expose an abuser as quickly as possible,” says Bours. “If we wait for the entire conversation to end, and the chatters have already made agreements, it could be too late. The monitor can also tell the child in the chat that they’re talking to an adult and not another child.”
AiBA has been collaborating with gaming companies to install the algorithm and is working with a Danish game and chat platform called MoviestarPlanet, which is aimed at children and has 100 million players.
In developing the algorithms, the researcher found that users write differently on different platforms such as Snapchat and TikTok.
“We have to take these distinctions into account when we train the algorithm. The same with language. The service has to be developed for all types of language,” says Bours.
Looking at chat patterns
Most recently, Bours and his colleagues have been looking at chat patterns to see which patterns deviate from what would be considered normal.
“We have analysed the chat patterns —instead of texts — from 2.5 million chats, and have been able to find multiple cases of grooming that would not have been detected otherwise,” Bours says
“This initial research looked at the data in retrospect, but currently we are investigating how we can use this in a system that follows such chat patterns directly and can make immediate decisions to report a user to a moderator,” he says.
AiBA developed its algorithm with financial support from NTNU Discovery, NFR FORNY and NTNU’s Department of Information Security and Communication Technology. NTNU Technology Transfer has managed the IPR and commercialisation work.
Read more content from NTNU:
-
Artificial nose can sniff out damaged fruit and spoiled meat
-
Want to be the best? It’s all in your mindset
-
Millions of Earth's species remain unknown to us. Can we still identify them?
-
AI expert: We should not have AI that exploits people's weaknesses
-
Afghanistan veterans are more prone to anger
-
Could a digital driving instructor offer the same quality as a human, but at a lower cost?