Artificial Intelligence reveals three trends in South African domestic abuse

Digital abuse, infidelity, and alcohol abuse are emerging as common conversation topics between victims of domestic violence in South Africa and rAInbow, an artificial intelligence-powered smart companion.

Developed with funding partner, Sage Foundation, and social justice organisation, The Soul City InstituterAInbow allows users to ‘chat’ to a non-human over Facebook Messenger. It provides a safe space for domestic violence victims to access information about their rights, support options, and where they can find help – in friendly, simple language.

When we launched rAInbow in November last year, we didn’t expect that it would facilitate over 200,000 conversations with 7,000 users – 150,000 of those within the first three months of launch. One of the reasons we believe Artificial Intelligence (AI) can fill a gap in victim support is because many victims are uncomfortable talking to another person about their experience – due largely to social and cultural taboos, embarrassment, and shame.

The data gathered from anonymised rAInbow conversations** provides invaluable insight into this complex issue; insight that we can use to improve our communication and prevention strategies.

Digital abuse: Behind the screens

Around 30% of rAInbow users believe it’s acceptable for their partners to check their phones and to insist on knowing who they’re talking to at all times.

Yet this constitutes a form of verbal and/or emotional abuse because abusers exploit technology and social media to monitor, control, shame, stalk, harass, and intimidate their victims. In conversations with rAInbow, many victims reveal that they don’t know what constitutes digital abuse because they can’t recognise the signs.

You could be a victim of digital abuse if your partner demands to know your passwords and who you’re talking to, reads your messages, and dictates who you can be friends with on social media.

The bottom line is, when you’re in a relationship, all communication with your partner – be it digital or face-to-face – should be respectful. You should never feel pressured into doing anything you’re uncomfortable with.

Infidelity: Is cheating really abuse?

Infidelity emerged as one of the main challenges facing rAInbow users in abusive relationships. In such cases, the cheating partner usually blames you for his/her cheating, does it intentionally to hurt you, or threatens to cheat again to control you. Infidelity is often accompanied by lying, manipulation, and blame-shifting – all recognised abusive behaviours.

Technology has exacerbated the problem. It’s now easier to access dating sites, pornography, and chat platforms, facilitating behaviour like ‘sexting’, which some people may consider infidelity.

‘Alcohol made me do it’

Alcohol and drugs are common triggers for violent episodes, with rAInbow users saying their partners were more likely to lash out at them verbally or physically after they’d been drinking. While alcohol itself doesn’t cause domestic violence, it can aggravate already tense situations.

Alcohol impairs people’s judgement and behaviour, to the point where they may lose control and become aggressive, short-tempered, and abusive. In most situations, the abusive partner will blame the alcohol for their actions and may not remember what they did or said the next day. The abused partner, however, has to live with the memories and after-effects of the abuse.

Everyone’s responsibility

In his State of the Nation Address earlier this year, President Cyril Ramaphosa said violence against women and children has reached “epidemic proportions” and that ending abuse would be made an urgent national priority. Corporates, NGOs, and ordinary citizens also have a responsibility to end the scourge.

Technology like rAInbow provides the vital information needed to start driving radical change – at policy and societal level. The conversations that rAInbow is having with users is making us think differently about how to approach this issue. It’s apparent that we need targeted, personalised education drives that help victims identify abuse and explain how and where to get help. It’s also apparent that there’s a strong need for information that can be accessed in a safe, anonymous, and non-judgemental space.

We need to use the aggregated data that’s available to us to make better decisions about action plans and strategies. Solutions like rAInbow can provide governments with the information they need to tackle abuse.

To find out how you can contribute to the rAInbow project, e-mail humans@aiforgood.co.uk.

 ** All conversational data is anonymised. It is used to improve rAInbow and help organisations make better decisions about where to focus their efforts to combat abuse.