Daniel Gayo Avello
Full Professor at the University of Oviedo in the area of “Computer Languages and Systems”
The article analyzes the prevalence of hate speech and 'fake' accounts on Twitter (now X) following its purchase by Elon Musk.
To do so, data from early 2022 to mid-2023 was analyzed. Elon Musk's entrance with the sink into the Twitter offices took place on October 22, 2022 and graphically marked his takeover of the platform; consequently, the dataset more than reasonably covers the situation on the platform prior to the purchase and for more than enough time for Elon's changes to both the platform and among the staff, as well as his troll-in-chief attitude to 'permeate' among users.
The authors of the study found a significant increase in hate speech manifested as racism, homophobia and transphobia, along with increased user interaction with such content.
The study shows no reduction in the activity of fake accounts (such as bots), and they even point out that it may have increased.
The results contradict many of Elon Musk's claims regarding the reduction in 'fake' account activity after he joined the company and implemented his changes.
Overall, the study is solid, and the fact that Kristina Lerman signed it is a guarantee of quality, given the enormous limitations that exist for studying Twitter since its purchase by Elon Musk. The longitudinal approach followed, even if it does not allow us to make causality claims (i.e., we cannot 'accuse' Elon Musk of being to blame for the new problems at Twitter nor intentionally intend them), does clearly show that there is a change for the worse in the discourse and toxicity of the platform after the purchase.
The data capture was done with the Twitter API for Academic Research, which ensures systematic and transparent data collection within the limitations set by the company itself. It should be noted that academic access ceased around September 21, 2023; it is possible that this is one of the last extensive studies to be conducted with the academic API and that, as of today, a study of this nature is basically impossible to conduct.
As far as the methods used to detect hate speech or 'fake' accounts are concerned they are reasonable, in the sense that any investigator facing this type of investigation would use similar approaches. Of course, this does not mean that they are 'bulletproof'. Using a dictionary to find hate speech tweets and then running them through Perspective's API (a toxic content detection service) is reasonable, but not 100% accurate. That is, there is bound to be hate speech that has slipped through the cracks and there may be a percentage, surely small, of rude and profanity-laced texts that have been flagged as hate speech.
All in all, the approach is reasonable and I am sure that, in aggregate, the indicators of increased hate speech are correct.
As far as detection of 'fake' accounts is concerned, an approach based on detection of coordinated campaigns is most reasonable according to recent literature, rather than simplistically trying to determine whether or not an individual account is a bot.
The work is broadly consistent with previous research focused on the same question - namely, changes in Twitter's toxicity following Musk's purchase - in two ways:
- It shows an increase in hate speech (though not as extreme as, for example, Hickey et al. Auditing Elon Musk's impact on hate speech and bots). The authors of this study argue that the difference is because Hickey et al. included in their lexicon the word “retarded” which is now used as an insult by right-wing users but should not be considered (according to the authors) as hate speech toward any minority group.
- The activity of inauthentic accounts has not noticeably reduced and, in fact, may have increased (particularly among accounts posting about cryptocurrencies).
As far as interesting developments in this paper are concerned:
- The time period analyzed is much longer, both in the stage prior to the purchase by Elon Musk and in his stage as CEO of the company; studies of a social nature (and this one is not without reason) are more robust when the periods under study are longer.
- The study of inauthentic accounts uses different metrics for their detection, which offers greater robustness to the findings.
- Hate speech is not analyzed monolithically but in different dimensions (racism, homophobia and transphobia), finding that transphobic speech is the one that experienced the greatest growth (note: Elon Musk has made many comments that have been interpreted as transphobic).
- It openly confronts Elon Musk's claims with the researchers' findings.
- It takes into account external events that may have played some role in the changes in hate speech (e.g., a beer commercial starring a trans woman).
As far as implications are concerned:
- It allows to argue that having content moderation templates (Elon Musk fired most of them, I seem to recall) has a positive impact on the platform by reasonably keeping hate speech at bay.
- It empirically confirms the 'feelings' (and decisions) of many people and organizations that toxicity and hate on Twitter have increased and make it an unhealthy platform where you need to value being on or off.
- In the absence of other studies that focus on the increase in this type of discourse (racist, homophobic and transphobic) in other media (e.g., traditional press), in legislative chambers (e.g., US Congress or Senate proceedings) or even the increase in hate crimes, one can only make 'guesses' about the real-world impact of this increase. However, the rise of hate speech on social media can influence the public agenda and offline attitudes, which should raise concerns about the potential impact (and harm) on individuals from the targeted hate groups.
Regarding limitations:
- Only English-language speech was analyzed, so nothing can be said about the rise of hate speech or inauthentic activity in other cultures. The choice of language, however, is reasonable given that, with hindsight, the purchase of Twitter could be related to the U.S. election. In this regard, it should be noted that, although the majority of the tweets in English were probably from US citizens, users from other countries were also present.
- The analysis of hate speech focused on clearly hateful speech (toxicity according to the Perspective API greater than or equal to 0.7); the authors themselves point out that the study of more subtle/sibylline hate speech remains to be done.
- The study of coordinated and inauthentic activity could be refined.
- The study relies heavily on third-party tools such as the Perspective API or Botometer (which has been openly criticized, see here).
- Apparently, the authors already saw changes and limitations in academic access during the conduct of their study (see page 6 of the article).
However, I must insist that many of the limitations are not inherent to this study but to any study conducted on Twitter and are consequently unavoidable.
All in all, I think it is a robust, interesting study that makes a strong case about hate speech and toxicity on Twitter in the U.S. post-Elon Musk buyout.