Keep Factually independent

Whether you agree or disagree with our analysis, these conversations matter for democracy. We don't take money from political groups - even a $5 donation helps us keep it that way.

Loading...Time left: ...
Loading...Goal: $500

Fact check: Is Twitter's algorithm actually visible for all users to see? If yes, does it mean we can see it's promoting right wing content?

Checked on February 7, 2025

1. Summary of the results

Twitter's algorithm is partially visible to the public. The company released its Recommendation Algorithm source code on GitHub on March 31, 2023 [1], including components related to the For You Timeline and Recommended Notifications [2]. However, it's important to note that while the code is public, Twitter did not release the training data or model weights [1], which means we can only see the basic mechanics but not the full implementation.

Regarding right-wing content promotion, multiple studies have confirmed this bias. A 2021 study demonstrated that Twitter's content recommendation system amplified right-leaning posts across multiple countries, including Canada, France, Japan, Spain, the UK, and the United States [3]. This issue has become serious enough that French prosecutors have launched an investigation into algorithmic distortions on the platform [4].

2. Missing context/alternative viewpoints

The algorithm's visibility is more complex than a simple yes/no answer. While the code is public, understanding it requires:

  • Knowledge of how the algorithm ranks content based on multiple factors including personal interest, location, recency, and virality [5]
  • Understanding of user engagement patterns, interests, and interaction history [6]
  • Recognition that this transparency initiative was part of Elon Musk's "new era of transparency" [1]

3. Potential misinformation/bias in the original statement

The original question oversimplifies two complex issues:

1. Algorithm Transparency: While Elon Musk and Twitter promote this as full transparency [1], the absence of training data and model weights means we can only see the framework, not the actual decision-making process.

2. Political Bias: While studies have shown right-wing content amplification [7], it's important to consider who benefits from these findings:

  • Platform owners: Elon Musk benefits from the perception of transparency by suggesting third-party analysis [1]
  • Political actors: Both right and left-wing groups can use these findings to support their narratives about platform bias
  • Researchers and regulators: Academic institutions and government bodies (like French prosecutors) gain influence through investigating these issues [4]
Want to dive deeper?
Jamal Roberts gave away his winnings to an elementary school.
Did a theater ceiling really collapse in the filming of the latest Final Destination?
Is Rachel Zegler suing South Park?