Insights AI News AI tool reducing polarization on X shows how to cool feeds
post

AI News

29 Nov 2025

Read 15 min

AI tool reducing polarization on X shows how to cool feeds

AI tool reducing polarization on X reorders feeds to lower partisan animosity and calm user reactions

A Stanford-led study shows an AI tool reducing polarization on X can cool feeds by downranking posts with anti-democratic attacks and partisan insults. The web extension uses a large language model to detect toxic content, then reorders the timeline. In tests, users felt warmer toward the other party and less angry. Social media can push hot content to the top. Angry posts get more clicks. Then the feed grows harsher, and people grow apart. A new research tool takes a softer path. It does not delete posts. It simply moves harsh political posts farther down your feed. This small change helped people feel less hostile toward the other party. The effect held for users from both sides. The study appears in Science and comes from a team including Stanford researchers. The tool works as a browser-based layer over X. It does not need help from the platform. That matters. It shows users and researchers can test and shape ranking choices without a company’s direct support. The team also shared code, so others can build on it and check the methods.

How the AI tool reducing polarization on X works

Reorder, don’t remove

Most feeds rank content to boost engagement. That often lifts posts that spark fear or anger. The research tool takes a different approach. It uses a large language model to scan posts in your timeline. It looks for signs of anti-democratic ideas, calls for political violence, or insults aimed at the other party. When it sees such posts, it does not hide them. It moves them lower, so you scroll past them later, if at all. This is key. The tool keeps speech intact. It changes the order to reduce constant exposure to the most heated content. That is less intrusive than bans or blocks. It also helps avoid backlash that comes with removal.

What the model flags

The classifier focuses on clear risk signals, such as:
  • Language that supports political violence or threats
  • Messages that push extreme measures against an opposing group
  • Content that undermines democratic norms or basic rights
  • Hostile partisan slurs that aim to dehumanize or humiliate
  • The goal is not to police ideas. It is to tone down repeated exposure to posts that fuel anger and distrust. The rest of the feed stays intact.

    An independent layer over a closed system

    Most social platforms run closed algorithms. Outsiders cannot see or adjust them. This tool shows a path around that barrier. As a web extension, it reads the feed you see and reorders it locally in your browser. It gives users and researchers a way to test alternative rankings without getting permission from the platform. That unlocks more transparency and more experimentation.

    The experiment and what changed

    Design and scale

    The team ran a 10-day study during the 2024 election period. About 1,200 people took part. Some saw their feeds as usual. Others used the extension that downranked toxic political content. The researchers measured feelings toward the other party and tracked emotions like anger and sadness.

    Results in plain numbers

  • Warmer attitudes: On a 100-point scale, views of the other party improved by about 2 points for users with downranked feeds.
  • Emotional relief: Reported anger and sadness fell for these users, compared to the control group.
  • Across the aisle: The effect was similar for liberal and conservative participants.
  • Big impact from a small nudge: The shift matched changes that usually happen in the wider population over several years, not days.
  • In short, the gentle reordering produced real, measurable change. It did not depend on telling people what to believe. It simply changed what appeared at the top of the feed.

    Why a small ranking change matters

    Ranking sets the tone. Most people do not scroll far. If your top posts attack the other side, your mind absorbs that mood. Over days, that mood can harden. By moving a few harsh posts down, the extension reduces daily blasts of contempt. This AI tool reducing polarization on X shows that tone and placement can shape how people feel about each other, even when nothing is deleted. These nudges are also less likely to spark fights over censorship. People keep access to all content. They just see the most inflammatory items later. That cuts temptations to doomscroll and lowers the chance of emotional spirals.

    User choice and transparent research

    Control for the person who scrolls

    People want a say in what they see. A simple switch to “cool the feed” offers that choice. It respects user goals. Some may use it during elections. Others may use it when they feel stressed. Giving people control is a practical path to healthier feeds.

    Open methods build trust

    The team made the code available. This allows other researchers and developers to test the approach, check for bias, and improve it. Open methods build trust and help the field move forward. It also lets civil groups and journalists review the rules for flagging content.

    Guardrails, bias, and accountability

    Any model can make mistakes. A fair tool needs guardrails:
  • Clear definitions: The tool should target specific harms, like threats or calls for violence, not broad political ideas.
  • Appeal options: Users should be able to flag false positives and adjust sensitivity.
  • Regular audits: Independent teams should test the model across parties, regions, and demographics.
  • Diverse data: Training and evaluation sets must include different dialects and contexts to reduce cultural bias.
  • Human oversight: Experts should review the rules and update them as norms and tactics shift.
  • Good guardrails reduce the risk of silencing a viewpoint by mistake. They also help keep the tool aligned with free expression.

    How teams can deploy similar approaches

    Here is a practical checklist for developers and researchers:
  • Define the harm: Focus on clear risk, like threats, dehumanizing slurs, or anti-democratic calls.
  • Start with a lightweight classifier: Use an LLM or a fine-tuned model that is fast and cost-aware.
  • Pick a ranking policy: Downrank flagged posts by a consistent amount rather than hiding them.
  • Measure user outcomes: Track changes in affect (anger, sadness), cross-party warmth, and trust in the feed.
  • Protect privacy: Run as much as possible locally in the browser. Minimize data sent to servers.
  • Get consent: Explain what the tool does and let users opt in and out easily.
  • Audit and iterate: Report error rates. Invite outside review. Update the model over time.
  • These steps promote safe, transparent, and effective feed interventions.

    Limits and open questions

    Longevity of the effect

    Does the warmer feeling last beyond ten days? We need longer studies to see if the change sticks and whether people adjust to the new order.

    Adversarial behavior

    Bad actors might change tactics to evade flags. Models will need updates, and tools should monitor for new patterns.

    Generalization to other platforms

    The method should work on any feed, but each platform has its own norms and layouts. More tests on different sites and in different languages are needed.

    Nuance and context

    Political speech is often charged. The line between tough criticism and dehumanizing attacks can be thin. Good design relies on narrow targets and human review of hard cases.

    Policy and platform implications

    What platforms can do now

    Platforms could offer optional “cool the feed” modes. They could expose ranking levers to users, such as “downrank content that attacks groups” or “reduce political rage-bait.” Even a simple slider would help. The study signals that small, optional changes can have big civic benefits.

    What regulators can support

    Policymakers can encourage:
  • User choice over ranking options
  • Transparency reports for recommender systems
  • Independent audits of civic-impact tools
  • Data access for public-interest research, with privacy protection
  • Such steps improve accountability without dictating speech.

    What civil society can add

    Civic groups and journalists can test the tool in real events. They can share findings and watch for bias. Schools and libraries could include these tools in media literacy programs, so people learn to manage their feeds. The AI tool reducing polarization on X is a model for collaboration. It shows how research, open code, and user control can align to improve public talk.

    What this means for elections and civic trust

    Democracy needs debate. It does not need daily doses of contempt. The study shows that exposure to fewer attacks can make people feel a bit warmer toward those who disagree. That matters in close elections, tense news cycles, and local disputes. The change is modest but real. It can lower the odds of spirals that push people to tune out or harden their views. This approach does not aim to change minds on issues. It aims to protect dignity. When your feed has fewer insults and threats, you have more room to think. You might read a post from the other side instead of dismissing it. You might discuss, not shout. That is a good baseline for any community.

    Why this approach stands out

    It respects speech

    Nothing is deleted. People keep the right to post and read. The tool edits order, not content.

    It empowers users

    People can choose a calmer experience. They do not need to accept a one-size-fits-all ranking.

    It invites scrutiny

    Open code and clear metrics welcome audits and improvements.

    It focuses on measurable outcomes

    The study connects a ranking change to specific improvements: warmer cross-party views and lower anger and sadness.

    Closing thoughts

    A healthier feed is not an abstract dream. It is a design choice. This research shows that a simple, visible shift—reordering posts that push violence or dehumanize others—can improve how we feel about people on the other side. It also shows that outsiders can test and refine ranking tools without waiting for a platform’s permission. With strong guardrails and open review, similar ideas could help many feeds, languages, and communities. For now, the key takeaway is clear: small, respectful changes to what rises to the top can cool online talk. The AI tool reducing polarization on X offers a practical, user-first path to make political conversation calmer, fairer, and more humane—at election time and every day.

    (Source: https://www.openaccessgovernment.org/new-ai-tool-can-lower-political-temperature-and-partisan-rhetoric-through-algorithm-control/201773/)

    For more news: Click Here

    FAQ

    Q: What is the AI tool introduced by the Stanford study? A: The Stanford team developed a web-based browser extension that uses a large language model to scan X feeds for anti-democratic language and partisan attacks, then reorders the timeline to downrank those posts without deleting them. The AI tool reducing polarization on X works locally in the browser and does not require the platform’s cooperation. Q: How does the tool identify harmful or polarising posts? A: It uses an LLM-based classifier to flag clear risk signals such as calls for political violence, extreme measures against opposing groups, content that undermines democratic norms, and hostile partisan slurs. Flagged posts are downranked rather than removed, so the rest of the feed remains intact. Q: Does the tool delete or censor content from users? A: The AI tool reducing polarization on X does not delete or hide posts; it keeps speech intact and simply moves incendiary posts lower in the feed so users encounter them later, if at all. This approach aims to reduce exposure to heated content while avoiding the backlash associated with removals. Q: What were the main findings from the user experiment? A: In a 10-day experiment during the 2024 election with roughly 1,200 participants, those who saw downranked toxic content reported about a 2-point warmer view of the opposing party on a 100-point scale and lower feelings of anger and sadness. The effect was observed across both liberal and conservative users. Q: Does the tool need cooperation from X or other platforms to function? A: No, the tool runs as an independent browser-layer that reads the feed a user already sees and reorders it locally in the browser, so it does not need platform permission to operate. The researchers also shared the code and methods so others can test alternative ranking policies without relying on the platform. Q: What guardrails are recommended to prevent bias or mistakes? A: The study recommends clear harm definitions, user appeal options and sensitivity controls, regular independent audits, diverse training and evaluation data, and human oversight to review difficult cases. These guardrails aim to reduce false positives and keep the tool aligned with free expression. Q: Can this approach be applied to other social media platforms and languages? A: The article suggests the method should generalize in principle but stresses that each platform has different norms and layouts, so more tests are needed across sites and languages. Additional research is required to validate effectiveness and adjust models for different contexts. Q: What policy or platform changes could follow from this research? A: Platforms could offer optional “cool the feed” modes or expose ranking levers like sliders to downrank content that attacks groups, while policymakers could promote user choice, transparency reports for recommender systems, independent audits, and controlled data access for public-interest research. These steps aim to improve accountability without mandating content removal.

    Contents