Published: June 11, 2018 By
Smart phone

Call them the cyberbully detectives: Researchers at CU 麻豆影院听have designed a new technique for spotting nasty personal attacks on social media networks like Instagram.

The new approach, developed by the听, is a response to a pernicious online problem that has been linked to the suicides of multiple teenagers in recent years. The method combines several different computing tools to scan massive amounts of social media data, sending alerts to parents or network administrators that abuse has occurred.听

It鈥檚 rocket fast, too: In , the group reported that their approach uses five times less computing resources than existing tools. That鈥檚 efficient enough to monitor a network the size of Instagram for a modest investment in server power, said study co-author Richard Han.

He and his colleagues hope that social media companies will take note, implementing the team鈥檚 toolset in their own networks. Han compared the prevalence of cyberbullying to the spread of fake news on social media, which Facebook CEO Mark Zuckerberg spoke to European leaders about in May. 听

BullyAlert screenshotBullyAlert screenshot

BullyAlert app notification screenshots.

鈥淭he response of the social media networks to fake news has recently started to uptick, even though it took grave consequences to reach that point,鈥 said Han, an associate professor in the Department of Computer Science. 鈥淭he response needs to be just as strong for cyberbullying.鈥

The group also released a that allows parents to receive alerts when their kids are the objects of bullying on Instagram. Study co-author Shivakant Mishra said that the app, called BullyAlert, is a valuable tool because it can learn from and adapt to what parents consider bullying.听

鈥淎s a parent, I know that a lot of times we are not in full knowledge of what our children are doing on their social networks,鈥 said Mishra, a professor in Computer Science. 鈥淎n app like this that informs us when something problematic is happening is invaluable.鈥

Building these sorts of resources, however, is a technical challenge鈥攁kin to finding a toxic needle in a very large haystack.听

Mishra explained that users have posted 40 billion photos and videos to Instagram since its launch in 2010. As a result, 鈥渨e want to be able to process as many posts as possible as fast as possible,鈥 he said. 鈥淏ut at the same time, we don鈥檛 want our accuracy to go down.鈥

To keep that balance, the CyberSafety team first employed real humans to teach a computer program how to separate benign online comments (鈥淟ove you!鈥) from abuse (鈥淵ou suck.鈥)

Next, the researchers designed a system that works a bit like hospital triage. When a user uploads a new post鈥攕ay, a photo of a new pasta recipe鈥攖he group鈥檚 tools make a quick scan of the comments. If those comments look questionable, then that post gets high priority to receive further checks. But if the comments all seem charitable, then the system bumps the post to the bottom of its queue.听

鈥淥ur goal is to focus on the most vulnerable sessions,鈥 Han said. 鈥淲e still continue to monitor all of the sessions, but we monitor more frequently those sessions that we think are more problematic.鈥

And it works: The researchers tested their approach on real-world data from Vine, a now-defunct video-sharing platform, and Instagram, which is especially popular among younger Internet surfers. Han explained that the team picked those networks because they make their data publicly available.

In research presented at a conference in April, the group calculated that their toolset could monitor traffic on Vine and Instagram in real-time, detecting cyberbullying behavior with 70 percent accuracy. What鈥檚 more, the approach could also send up warning flags within two hours after the onset of abuse鈥攁 performance unmatched by currently available software.听

Han said that social media companies have no excuse not to implement such approaches for monitoring cyberbullying. He explained that online harassment can be more dangerous than schoolyard bullying because it can reach kids anywhere, even in their own homes.听

The message to social media companies is that 鈥減eople are abusing your networks to spread fake news, but also they鈥檙e also hurting kids,鈥 Han said. 鈥淎ll of us need to do better to protect the kids.鈥

Rahat Ibn Rafiq, a CU 麻豆影院 graduate student in Computer Science, is the lead author of the new study. Other co-authors include CU 麻豆影院 Associate Professor Qin (Christine) Lv, and Homa Hosseinmardi of the University of Southern California.