Ethereum co-founder Vitalik Buterin called on X to use blockchain and ZK-proofs to make its content-ranking algorithm verifiably fair and censorship-resistant.

News
Ethereum co-founder Vitalik Buterin is calling for social media platforms to use cryptography and blockchain tools to make their content-ranking systems more transparent and verifiable.
In a Monday X , Buterin argued that X should use zero-knowledge proofs (ZK-proofs) and blockchain to prove the fairness of the algorithm that determines the reach of content on the platform. He raised an issue with how X operates on Tuesday, claiming that owner Elon Musk’s way of leading it was harmful:
“Elon Musk I think you should consider that making X a global totem pole for free speech, and then turning it into a death star laser for coordinated hate sessions, is actually harmful for the cause of free speech.“
Ethereum Foundation AI lead Davide Crapis reacted to this initial idea by saying, “If you want to claim X is the platform for free speech, you should disclose your algorithm optimization targets.” He added that “it should be legible to the users, and tweakable.”
Buterin suggested a verifiable system that employs ZK-proofs for every decision made by the algorithm and timestamps all content, likes, and retweets on a blockchain “so the server can’t censor or lie about time.” The platform should also “commit to publishing the full algorithm code with a 1-2 year delay.”

ZK-proofs are a cryptographic way to prove that something is true without revealing the underlying data — for instance, proving that you are over 18 without sharing your full name. Buterin did not go into detail about what proofs would demonstrate in his suggested solution, but they would likely show that algorithmic decisions followed certain constraints without sharing sensitive details.
Related:
Crypto takes on social media
Buterin’s post echoed the sentiment behind some . Such initiatives, despite none having reached mainstream adoption, appear to be taken seriously by their traditional centralized equivalents.
In early 2025, Meta, the parent company of Facebook and Instagram, called Pixelfed. All links to the platform were labeled as “spam” and removed immediately. Others claimed that Facebook competitors, including Mastodon, were given the same treatment.
The crypto community — with its tendency to be wary of centralized control — has raised concerns about the potential impact of decisions made by social media platform leadership in the past. When Musk announced in early January that X would prioritize promoting over other types of content, many were not convinced.
Critics questioned who would decide what qualifies and whether the policy could become a vehicle for suppressing certain viewpoints. Musk has also been accused by critics of limiting access to premium features for users who disagreed with him.
Buterin also chimed in at the time. He urged Musk to on the platform and not ban users for disagreements or expressing views.
Related:
The impact of social media on society
Research has long shown that social media has an outsized impact on society and the functioning of democratic processes. A paper in 2024 suggested that “access to Facebook may increase belief in misinformation.”
Reuters also reported last month that recent court filings that Meta shut down internal research into the mental health effects of Facebook after finding causal evidence that its products harmed users’ mental health. The study found that “people who stopped using Facebook for a week reported lower feelings of depression, anxiety, loneliness, and social comparison.”
The European Union to tackle the issue with its Digital Services Act, which requires transparency on the main algorithm parameters and requires platforms to assess the risk and disclose the findings on the potential negative impact of their operations. Impacts considered explicitly include “negative effects on civic discourse and electoral processes, and public security.”
The Digital Services Act also requires that vetted researchers be given access to platform data to study their systemic risk independently. A lack of compliance by X with this specific requirement is among the reasons cited by the European Commission for a 120 million euro ($141 million) fine earlier this month.
Other reasons include a lack of transparency on X’s ads repository and the platform’s blue checkmark purportedly deceiving users since “anyone can pay to obtain the ‘verified’ status without the company meaningfully verifying who is behind the account.”






















