{"id":55417,"date":"2023-10-16T19:54:59","date_gmt":"2023-10-16T23:54:59","guid":{"rendered":"https:\/\/euvolution.com\/open-source-convergence\/uncategorized\/does-deplatforming-work-to-curb-hate-speech-and-calls-for.php"},"modified":"2023-10-16T19:54:59","modified_gmt":"2023-10-16T23:54:59","slug":"does-deplatforming-work-to-curb-hate-speech-and-calls-for","status":"publish","type":"post","link":"https:\/\/euvolution.com\/open-source-convergence\/deplatforming\/does-deplatforming-work-to-curb-hate-speech-and-calls-for.php","title":{"rendered":"Does &#8216;deplatforming&#8217; work to curb hate speech and calls for &#8230;"},"content":{"rendered":"<p><p>    In the wake of the     assault on the U.S. Capitol on Jan. 6, Twitter     permanently suspended Donald Trumps personal account, and        Google, Apple and Amazon shunned Parler, which at least    temporarily shut down the social media platform favored by the    far right.  <\/p>\n<p>    Dubbed deplatforming, these actions     restrict the ability of individuals and communities to    communicate with each other and the public. Deplatforming        raises ethical and legal questions, but foremost is the    question of whether its an effective strategy to reduce hate    speech and calls for violence on social media.  <\/p>\n<p>    The Conversation U.S. asked three experts in online    communications whether deplatforming works and what happens    when technology companies attempt it.  <\/p>\n<p>    Jeremy Blackburn, assistant professor of computer    science, Binghamton University  <\/p>\n<p>    The question of how effective deplatforming is can be looked at    from two different angles: Does it work from a technical    standpoint, and does it have an effect on worrisome communities    themselves?  <\/p>\n<p>    Does deplatforming work from a technical perspective?  <\/p>\n<p>    Gab was the first major platform     subject to deplatforming efforts, first with removal from    app stores and, after the     Tree of Life shooting, the withdrawal of cloud    infrastructure providers, domain name providers and other    Web-related services. Before the shooting, my colleagues and I    showed in a study that Gab was an alt-right echo    chamber with worrisome trends of hateful content. Although    Gab was deplatformed, it managed to survive by shifting to    decentralized    technologies and has shown a degree of innovation  for    example, developing the moderation-circumventing    Dissenter browser.  <\/p>\n<p>    From a technical perspective, deplatforming just makes things a    bit harder. Amazons cloud services make it easy to manage    computing infrastructure but are ultimately built on open    source technologies available to anyone. A deplatformed company    or people sympathetic to it could build their own hosting    infrastructure. The research community has also built     censorship-resistant tools that, if all else fails, harmful    online communities can use to persist.  <\/p>\n<p>    Does deplatforming have an effect on worrisome communities    themselves?  <\/p>\n<p>    Whether or not deplatforming has a social effect is a nuanced    question just now beginning to be addressed by the research    community. There is evidence that a platform banning    communities and content  for example, QAnon or certain    politicians  can have a positive effect. Platform banning can    reduce growth of new    users over time, and there is less content produced    overall. On the other hand, migrations do happen,    and this is often a response to real world events  for    example, a deplatformed personality who migrates to a new    platform can trigger    an influx of new users.  <\/p>\n<p>    Another consequence of deplatforming can be users in the    migrated community showing signs of becoming more radicalized    over time. While Reddit or Twitter might improve with the    loss of problematic users, deplatforming can have unintended    consequences that can accelerate the problematic behavior that    led to deplatforming in the first place.  <\/p>\n<p>    Ultimately, its unlikely that deplatforming, while certainly    easy to implement and effective to some extent, will be a    long-term solution in and of itself. Moving forward, effective    approaches will need to take into account the complicated    technological and social consequences of addressing the root    problem of extremist and violent Web communities.  <\/p>\n<p>    Ugochukwu Etudo, assistant professor of operations and    information management, University of Connecticut  <\/p>\n<p>    Does the deplatforming of prominent figures and movement    leaders who command large followings online work? That depends    on the criteria for the success of the policy intervention. If    it means punishing the target of the deplatforming so they pay    some price, then without a doubt it works. For example,    right-wing provocateur Milo Yiannopoulos was banned from    Twitter in 2016 and Facebook in 2019, and subsequently     complained about financial hardship.  <\/p>\n<p>    If it means dampening the odds of undesirable social outcomes    and unrest, then in the short term, yes. But it is not at all    certain in the long term. In the short term, deplatforming    serves as a shock or disorienting perturbation to a network of    people who are being influenced by the target of the    deplatforming. This disorientation can weaken the movement, at    least initially.  <\/p>\n<p>    However, there is a risk that deplatforming can delegitimize    authoritative sources of information in the eyes of a    movements followers, and remaining adherents    can become even more ardent. Movement leaders can reframe    deplatforming as censorship and further proof of a mainstream    bias.  <\/p>\n<p>    There is reason to be concerned about the possibility that    driving people who engage in harmful online behavior into the    shadows further entrenches them in online environments that    affirm their biases. Far-right groups and personalities have    established a considerable presence on privacy-focused online    platforms, including the messaging platform    Telegram. This migration is concerning because researchers    have known for some time that complete online anonymity is    associated with increased    harmful behavior online.  <\/p>\n<p>    In deplatforming policymaking, among other considerations,    there should be an emphasis on justice, harm reduction and    rehabilitation. Policy objectives should be defined    transparently and with reasonable expectations in order to    avoid some of these negative unintended consequences.  <\/p>\n<p>    Robert Gehl, associate professor of communication and    media studies, Louisiana Tech University  <\/p>\n<p>    Deplatforming not only works, I believe it needs to be built    into the system. Social media should have mechanisms by which    racist, fascist, misogynist or transphobic speakers are    removed, where misinformation is removed, and where there is no    way to pay to have your messages amplified. And the decision to    deplatform someone should be decided as close to democratically    as is possible, rather than in some closed boardroom or opaque    content moderation committee like     Facebooks Supreme Court.  <\/p>\n<p>    In other words, the answer is alternative social media like    Mastodon. As a    federated system, Mastodon is specifically designed to give    users and administrators the ability to mute, block or even    remove not just misbehaving users but entire parts of the    network.  <\/p>\n<p>    For example, despite fears that the alt-right network Gab would    somehow take over the Mastodon federation,     Mastodon administrators quickly marginalized Gab. The same    thing is happening as I write with new racist and misogynistic    networks forming to fill the potential void left by Parler. And    Mastodon nodes have also prevented spam and advertising from    spreading across the network.  <\/p>\n<p>    Moreover, the decision to block parts of the network arent    made in secret. Theyre done by local administrators, who    announce their decisions publicly and are answerable to the    members of their node in the network. Im on scholar.social, an    academic-oriented Mastodon node, and if I dont like a decision    the local administrator makes, I can contact the administrator    directly and discuss it. There are other distributed social    media system, as well, including Diaspora and Twister.  <\/p>\n<p>    The danger of mainstream, corporate social media is that it was    built to do exactly the opposite of what alternatives like    Mastodon do: grow at all costs, including the cost of harming    democratic deliberation. Its not just cute cats that draw    attention but conspiracy theories, misinformation and the    stoking of bigotry. Corporate social media tolerates these    things as long as theyre profitable  and, it turns out, that    tolerance has lasted far too long.  <\/p>\n<p>    [Deep knowledge, daily.     Sign up for The Conversations newsletter.]  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>See the original post here:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/theconversation.com\/does-deplatforming-work-to-curb-hate-speech-and-calls-for-violence-3-experts-in-online-communications-weigh-in-153177\" title=\"Does 'deplatforming' work to curb hate speech and calls for ...\">Does 'deplatforming' work to curb hate speech and calls for ...<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> In the wake of the assault on the U.S. Capitol on Jan. 6, Twitter permanently suspended Donald Trumps personal account, and Google, Apple and Amazon shunned Parler, which at least temporarily shut down the social media platform favored by the far right<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[60423],"tags":[],"class_list":["post-55417","post","type-post","status-publish","format-standard","hentry","category-deplatforming"],"_links":{"self":[{"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/posts\/55417"}],"collection":[{"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/comments?post=55417"}],"version-history":[{"count":0,"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/posts\/55417\/revisions"}],"wp:attachment":[{"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/media?parent=55417"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/categories?post=55417"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/euvolution.com\/open-source-convergence\/wp-json\/wp\/v2\/tags?post=55417"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}