Blackburn’s Quest: Computer science professor researches darkest corners of the internet | Binghamton News – Binghamton University

Life in the digital age has added one more certain thing to the old saying about death and taxes: People are going to be jerks on the internet.

Whether its an anonymous troll questioning your parentage or a propaganda campaign by a foreign power, the signal-to-noise ratio on social media has become much worse in recent years. Thats not even mentioning the hate-mongers, conspiracy theorists and outright liars who want their skewed views to become your views.

Assistant Professor Jeremy Blackburn, a faculty member in Watson Colleges Department of Computer Science, has been researching bad actors online for more than 10 years. That journey has taken him to some dark places where outsiders fear to tread, but he hopes that by shining a light there, we can start to figure how to fix them.

I dont think the problems are new. They are fundamental human problems, Blackburn says. Whats different is that its become a socio-technical problem rather than just a social problem. The internet doesnt make people bad it just enables them to be worse, and it enables them to find other people who are also bad.

FROM GAMING TO SOCIAL MEDIA

Blackburn first became interested in computers while growing up in Florida, connecting with fellow users from around the world through massively multiplayer online role-playing games (MMORPGs) such as Ultima Online. Players adopted sword-and-sorcery character avatars for quests to conquer kingdoms and battle monsters.

Because Blackburn and his friends were clever with programming code, they sometimes would find ways to cause chaos. One time, his clan built a virtual house in front of a key entry point and shot arrows from inside at other players who approached. Another trick, which landed them in the games jail, involved killing a character and stealing the blueprints for a new kind of building being beta-tested.

Yeah, they werent exactly angels.

If you did that kind of stuff in person during a Dungeons & Dragons game, you might get punched in the mouth, Blackburn says with a laugh. But the fact that it was virtual enabled a whole different level of mischief.

Like many teens who love coding, Blackburn headed to college in his case, the University of South Florida (USF) in Tampa with the intent to design computer games. His interests later shifted to the underlying technologies that make shared games possible, such as distributed systems that spread various components across multiple computers.

For his doctoral thesis also at USF he returned to the idea of bad behavior online by studying cheating in internet gaming, and that drew a direct path to the kind of research he does today.

While earning his degrees, Blackburn worked for more than a decade in private industry, including as principal developer at test-prep company Boson Software and as software architect at his own company, Pallasoft. He also spent three years as an associate researcher at Telefonica Research in Barcelona, Spain.

His time in academia first at the University of Alabama at Birmingham and now at Binghamton University has coincided with the proliferation and influence of mainstream platforms such as Facebook and Twitter as well as niche apps like Telegram, Parler, 4chan and Gab.

Things have evolved away from blogs and similar sites in the past 10 years, he says. People want interactive social media they want to be able to engage with each other rather than just scream on a soapbox.

In our polarized society, though, those back-and-forth interactions can get downright nasty.

TRACKING THE TROLLS

Blackburn is the co-founder of the International Data-driven Research for Advanced Modeling and Analysis (iDRAMA) Lab, which includes more than two dozen professors, PhD students and industry researchers from around the world.

In various configurations, iDRAMA members have studied nearly every social media platform, from dominant ones like Twitter to white supremacist havens such as Gab and 4chan. The only one they ignore is Facebook, because data collection from there has become increasingly unreliable.

Recent research

Assistant Professor Jeremy Blackburn has studied many types of bad actors on the internet. Here are just a few of the findings from the iDRAMA Lab in recent years.

CYBERBULLYING: By analyzing the behavioral patterns of abusive Twitter users and their differences from other Twitter users, Blackburn and his colleagues developed machine learning algorithms that can successfully identify bullies and aggressors on Twitter with 90% accuracy.

ZOOMBOMBING: As more companies and schools moved to online platforms like Zoom, Google Meet and Skype in the early days of the COVID-19 pandemic, interruptions to those meetings and classes often racist or pornographic became an increasing problem. An iDRAMA study showed that most of those attacks were not due to hackers, but because legitimate attendees gave passwords and advice to friends or posted the information on Reddit, Twitter and 4chan.

DEPLATFORMING: Once users are banned from Twitter and Reddit, what happens to them? Often they move to smaller platforms like Gab or Parler, where the content moderation is more lax. There, they have a potentially reduced audience but exhibit an increased level of activity and toxicity than they did previously.

ANTI-ASIAN HATE SPEECH: By analyzing posts from 4chans Politically Incorrect board as well as Twitter from late 2019 and early 2020, the iDRAMA lab tracked the rise of Sinophobic content related to the pandemic.

The iDRAMA Lab has published research on QAnon, the rise in anti-Asian and anti-Semitic sentiments, the use of manipulated news images (also known as fauxtography), cyberbullying, misogyny, state-sponsored disinformation campaigns and more.

Its a roundup of the worst that humanity has to offer, and sometimes the haters strike back. A recent 4chan post, for instance, claimed that Blackburn is a Hamas recruiter, and hes received a few ominous threats over the years. (Luckily, nothing came of them.)

Blackburn fosters an atmosphere of camaraderie among his students and peers, welcoming open conversations so that no one feels overwhelmed by internet hate.

If you dont look at the content, you cant really do research about it, he says, but if you look at the content too much or too deeply if you stare into the abyss a bit too long you might fall into it. Its hard walking that line, and Ive certainly had failures along the way.

Gianluca Stringhini, an assistant professor at Boston University and co-founder of the iDRAMA Lab, praises Blackburns willingness to think outside of the boundaries of traditional computer science methods.

When Jeremy and I started working together, we realized that studying these emerging sociotechnical problems required techniques that dont really fall under any of the established research methods in our fields, Stringhini says.

Five years later, we are combining computer networks, security, graph analysis, psychology and other disciplines to paint a comprehensive picture of online weaponized information. Not many researchers would be comfortable doing that, but Jeremy has a unique vision and is not afraid of breaking with research norms.

TURNING OVER THE ROCKS

Earlier this year, Blackburn received a five-year, $517,484 National Science Foundation CAREER Award for his project Towards a Data-Driven Understanding of Online Sentiment. The CAREER Award supports faculty who have the potential to serve as future academic role models.

At the core of the project is devising a better way to train machine learning which does most of the content moderation on social media platforms about how to judge the offensiveness of images used in memes.

Currently, artificial intelligence software tries to determine if a particular image is bad or not, but Blackburn wants to take a trick from online gaming by presenting it two images and asking which is worse. The process is similar to the matchmaking system that puts gamers into groups of similar skills, not people who are 1,000 times better or worse than you.

Instead of looking at images in isolation and making a judgment on that individual piece of content, its more like ordering them, he says. Were not learning if something is racist or not were learning which is more racist. Who knows what well find, but were convinced that it will lead to something interesting.

Blackburn admits that he and his iDRAMA colleagues sometimes discuss whether their research is helping internet jerks to dig in deeper and evade future detection. Maybe if they didnt turn over the rocks, the nasty critters underneath would just stay there and never come out.

As a computer scientist, though, Blackburn believes that learning more will be an important step toward curbing what has become a political and social menace. He contends its also a public health crisis: Online hate affects our mental well-being, and misinformation about COVID-19 has led to more deaths and hospitalizations.

We have this insanely powerful, world-changing technology thats been around for less than a generation, he says. I hope that well provide the knowledge and tools to become more resilient, more robust and less susceptible to this type of behavior, and to start figuring out ways to actively address it.

The rest is here:

Blackburn's Quest: Computer science professor researches darkest corners of the internet | Binghamton News - Binghamton University

Related Posts
This entry was posted in $1$s. Bookmark the permalink.