veda.ng
Essays/Computational Social Science

Computational Social Science

Twitter's API shutdown disrupted 20 years of social research. Digital twins simulate voter behavior.

Vedang Vatsa·September 5, 2025·9 min read
Infographic
The Core Thesis

Social science was built on surveys, interviews, and small-sample experiments. Computational social science (CSS) replaces these with the behavioral traces that billions of people generate daily: posts, clicks, purchases, movements, searches. The instruments are different: machine learning instead of regression, network analysis instead of focus groups, platform data instead of questionnaires. The scale is different: millions of observations instead of hundreds. The constraints are also different: platform API access is gated by corporate interest, algorithmic filtering distorts observed behavior, and the populations studied are not representative. CSS is the most powerful tool social science has ever had, and the most fragile.

The Data Source Problem

For two decades, Twitter (now X) was the primary data source for computational social science. Its API provided researchers with near-real-time access to public posts, enabling studies of political polarization, misinformation spread, disaster response, public health sentiment, and social movement dynamics. An estimated 25,000+ peer-reviewed papers were built on Twitter data, making it the single most studied platform in the history of social science.

25,000+
Peer-reviewed papers built on Twitter data
Google Scholar estimates
$42K/mo
Cost of comparable X API access (post-2023)
23,000
Participants in Meta's 2023 Science studies
689,000
Users in Facebook's emotional contagion study
PNAS (2014)

In 2023, following Elon Musk's acquisition, Twitter imposed API pricing that effectively ended free academic access. The academic tier, which had provided researchers with up to 10 million tweets per month at no cost, was eliminated. The replacement pricing, approximately $42,000 per month for comparable access, was prohibitively expensive for most university research budgets. Reddit followed a similar path in mid-2023, shutting down the widely used Pushshift archive and introducing API pricing at $0.24 per 1,000 calls.

The consequences for the field are structural. Longitudinal studies that tracked political behavior, public health attitudes, or information diffusion over years lost their data stream overnight. Replication of prior findings became difficult or impossible. The platforms that had served as a public commons for social observation became private resources available primarily to well-funded corporate and government entities.

The disruption forced a reckoning with a dependency that the field had not adequately recognized: CSS had built its methodological infrastructure on the assumption that platform data access was a stable public good. It was, in fact, a private resource that could be revoked at any time, for any reason, by a single corporate decision.

The Data Access Crisis

Platform API access for academic researchers (2025)

Twitter/XRestricted
Cost: $42K/mo
Papers built on: 25,000+
Impact: Primary source lost for most researchers
Meta (FB/IG)Limited partnerships
Cost: Application-based
Papers built on: 15,000+
Impact: 2023 Science studies via Meta partnership
RedditRestricted (2023)
Cost: $0.24/1K API calls
Papers built on: 10,000+
Impact: Pushshift archive shut down
BlueskyOpen (AT Protocol)
Cost: Free
Papers built on: Emerging
Impact: Decentralized, firehose API available
MastodonOpen (ActivityPub)
Cost: Free
Papers built on: Growing
Impact: Federated, instance-level access
TikTokMinimal
Cost: Research API (limited)
Papers built on: 2,000+
Impact: Most studied via scraping, not API

Sources: ICWSM conference proceedings, Twitter/X developer portal, Reddit API documentation. Paper counts are estimates from Google Scholar.

The Methodological Evolution

The loss of easy API access forced rapid methodological innovation. The new methods that emerged are, in several ways, superior to the API scraping they replaced, though each introduces its own limitations.

Browser extension studies. Researchers at NYU, Princeton, and other institutions now recruit volunteer participants who install custom browser extensions. These extensions log the content participants actually see in their social media feeds, not just what they post, but what the algorithm surfaces. This provides ground truth about algorithmic exposure that API data never offered: the old approach captured what users said, but not what they were shown. The limitation is scale: browser extension studies typically involve thousands of participants rather than the millions available through APIs, and participants are self-selected.

Data donation. GDPR Article 15 and equivalent regulations give individuals the right to export their complete interaction history from any platform. Academic protocols now guide participants through downloading their data packages, which include likes, saves, dwell time, scroll behavior, search history, and algorithmic recommendations, and sharing them with researchers under informed consent. This produces richer individual-level data than API scraping ever provided, though at far smaller scale and with significant effort required from each participant.

Digital twins. Researchers are using Large Language Models to create synthetic replicas of user personas: "digital twins" trained on real user-generated content that simulate how individuals might respond to different stimuli. A 2024 Stanford study demonstrated that GPT-4-based agents could replicate individual survey responses with 85% accuracy when calibrated on a person's social media history. These synthetic agents can model scenarios (policy changes, platform design modifications, information campaigns) without exposing real users to experimental manipulation. The fundamental limitation is validation: how do you confirm that synthetic behavior accurately represents real behavior without the real behavioral data you no longer have access to?

Algorithmic auditing. Rather than studying user behavior, a growing body of research studies algorithm behavior. Researchers create controlled "sock puppet" accounts with different disclosed demographics (age, gender, political orientation, location) and measure divergence in algorithmic treatment: what content is recommended, what is suppressed, what advertisements are shown. This approach treats the algorithm itself as the subject of study, not users, and has produced some of the field's most important findings about systematic amplification and suppression patterns.

Platform partnerships. The 2023 Meta/Science collaboration, which produced four landmark papers on algorithmic polarization, represents the partnership model: researchers work with platform companies under NDA to access internal data at scale. The advantage is access to data of unparalleled richness (including behavioral signals like dwell time and scroll velocity that are never exposed through APIs). The disadvantage is corporate control: Meta retained the right to review findings before publication, and critics have noted that the research design, which focused on short-term experimental manipulation of individual features, was structured in a way that was unlikely to find large effects.

Methodological Evolution

How CSS adapted after the API shutdown

API Scraping (traditional)declining
2006–2023
Scale: Millions of posts
Privacy: Public data
Limitation: Platform-dependent, shut down
Browser Extension Studies
2020–present
Scale: Thousands of participants
Privacy: Informed consent
Limitation: Small sample, self-selection bias
Data Donation (GDPR Art. 15)
2021–present
Scale: Hundreds to thousands
Privacy: User-controlled export
Limitation: Very small scale, labor-intensive
LLM Digital Twins
2023–present
Scale: Unlimited synthetic agents
Privacy: No real users exposed
Limitation: Validity of synthetic behavior
Algorithmic Auditing
2019–present
Scale: Controlled experiments
Privacy: Researcher-created accounts
Limitation: Platform detection, ToS violations
Platform Partnerships
2018–present
Scale: Millions (platform-controlled)
Privacy: NDA-governed
Limitation: Corporate veto on findings

The most important finding in computational social science is not about users. It is about platforms. The algorithm is not a neutral conduit for human behavior. It is an active participant that shapes, filters, amplifies, and distorts the behavior it claims to merely reflect.

The Polarization Question

The largest-scale CSS studies have focused on political polarization, with results that challenge the popular narrative that "social media causes polarization."

The 2023 Meta/Science collaboration experimentally modified the Facebook and Instagram feeds of 23,000 users during the 2020 US election. The research produced four papers, each testing a different intervention:

Chronological feed experiment. One group saw chronologically ordered content instead of algorithmically ranked content. The algorithmic feed increased exposure to ideologically aligned content and decreased exposure to cross-cutting perspectives. However, switching to chronological feeds did not measurably reduce political polarization in attitudes. Users exposed to chronological feeds spent less time on the platform, clicked less, and reacted to fewer posts.

Reshare removal experiment. Removing reshared content from Facebook feeds significantly reduced the volume of political news and misinformation that users encountered. The reshare mechanism, which allows content to propagate through networks at low cognitive cost (one tap), is a primary vector for both political content amplification and misinformation spread. This finding was arguably the most actionable: reshares are a design choice, not a natural behavior, and their removal has measurable effects on information quality.

Like-minded source reduction. Reducing the proportion of content from like-minded sources decreased exposure to news from untrustworthy sources. Political news consumption is highly ideologically segregated, with users seeking out and engaging more with content that aligns with their pre-existing views. The echo chamber effect is real but algorithmically maintained, not solely user-driven.

These findings are nuanced and frequently misquoted. They do not show that algorithms "don't cause" polarization. They show that short-term experimental manipulation of a single platform feature does not rapidly change deeply held political attitudes that have been forming over years or decades. The long-term, cumulative effect of years of algorithmic curation, which these studies, by design, could not measure, remains the critical open question.

What the Research Actually Shows

Landmark CSS studies on algorithmic polarization

Meta/Science (2023): Chronological feedn=23,000

Algorithmic feed ↑ ideological sorting; chronological feed did not ↓ polarization

Short-term manipulation ≠ attitude change

Meta/Science (2023): Remove resharesn=23,000

Removing reshared content ↓ political news & misinformation exposure significantly

Reshare mechanism is the primary amplification vector

Meta/Science (2023): Like-minded sourcesn=23,000

Reducing like-minded content ↓ news from untrustworthy sources

Echo chambers are algorithmically maintained

Bail et al. (2018): Cross-cutting exposuren=1,220

Exposure to opposing views ↑ polarization (for Republicans)

More information ≠ less polarization

Allcott et al. (2020): Facebook deactivationn=2,743

4-week deactivation ↓ political knowledge, ↓ polarization slightly

Platform use maintains engagement in political discourse

Sources: Science (2023), PNAS (2018, 2020). Meta studies conducted during 2020 US election cycle.

The Representation Problem

A critical limitation of CSS is population representativeness. No social media platform constitutes a representative sample of any national population.

Twitter/X's user base skews younger, more male, more urban, more politically engaged, and more socioeconomically advantaged than the general population. Facebook's user base skews older. TikTok's skews dramatically younger (Gen Z and Gen Alpha). Reddit's skews male and tech-literate. LinkedIn's skews professional and affluent. Each platform selects for a different subset of the population, and the design of each platform shapes the behavior it observes.

"Twitter sentiment" is not public sentiment. It is the sentiment of people who express opinions on Twitter, filtered through an algorithm that amplifies emotionally charged content. Research that draws conclusions about "public opinion" from any single platform's data is, in practice, drawing conclusions about the opinions of a non-representative subset of the public, expressed in a format that the platform's design incentivizes, and filtered through an algorithm optimized for engagement rather than accuracy.

The observer effect operates at population scale. Social media platforms are not neutral observation environments. Their design incentivizes specific behaviors: brevity (character limits on Twitter), emotional intensity (engagement algorithms reward outrage over nuance), public performance (visibility metrics create incentives for performative rather than genuine expression), and conformity (social proof through likes and shares discourages minority viewpoints). Behavior observed on these platforms is behavior shaped by these platforms. The instrument of measurement alters what is being measured.

The Observer Effect at Platform Scale

How platform design distorts the behavior it measures

PlatformDesign BiasAlgorithmic BiasPopulation Bias
Twitter/XBrevity (280 chars), public performanceEngagement optimization → outrage amplificationYounger, male, urban, politically engaged
FacebookSocial graph, reshare mechanicsPredicted engagement → ideological sortingOlder, broader demographics, declining youth
TikTokShort video, low-friction creationWatch time optimization → sensationalismGen Z / Gen Alpha, global
RedditPseudonymous, community-gatedUpvote sorting → popularity biasMale-skewed, tech-literate, Western

"Twitter sentiment" is not public sentiment. It is the sentiment of people who express opinions on Twitter, filtered by an algorithm that amplifies emotionally charged content.

The Ethics Infrastructure

CSS raises ethical questions that traditional social science frameworks were not built to handle.

The Facebook emotional contagion study (2014) remains the defining ethical controversy. The study experimentally manipulated the emotional content in 689,000 users' news feeds without informed consent, demonstrating that emotional states are contagious through social networks: users exposed to more negative content posted more negatively themselves. The backlash was not about the finding but about the method: manipulating the emotional experiences of nearly 700,000 people at scale, without their knowledge or consent, under the legal cover of Terms of Service rather than Institutional Review Board (IRB) approval.

The ethical standards that have evolved since then remain fundamentally inconsistent:

Institutional Review Boards (IRBs) at universities now require explicit informed consent for social media experiments involving any form of manipulation or observation of identifiable individuals. A graduate student studying 500 tweets needs IRB approval. But IRB jurisdiction does not extend to corporate research. Platform companies can conduct internal experiments on their own users, at any scale, under their Terms of Service. Facebook alone runs thousands of A/B tests per day on its user base, each one a social experiment conducted without informed consent.

Data protection regulations (GDPR in Europe, CCPA in California) provide individuals with legal rights over their data, including the right to access, export, and delete it. These regulations have created the legal foundation for data donation research but do not address the fundamental power asymmetry: platforms possess behavioral data of unprecedented granularity, while researchers must negotiate access on the platform's terms.

The "do no harm" principle extends to publication effects. In polarized environments, even publishing research about group behavior can be weaponized. Findings about the online behavior of a political, ethnic, or religious group can be cited, often out of context, to justify discrimination, surveillance, or platform-level suppression. CSS researchers increasingly face decisions about whether publishing accurate findings serves the public interest or provides ammunition to bad actors.

The Asymmetry Problem

A Facebook data scientist can run an experiment on 10 million users with no external oversight. An academic researcher studying 500 public tweets requires months of IRB review. This asymmetry means that the most consequential social experiments in human history are being conducted by corporate employees, not independent researchers, and the results are proprietary.

The Emerging Infrastructure

The field is building toward a more resilient methodological infrastructure that does not depend on the goodwill of platform corporations.

Decentralized platforms. Bluesky (built on the AT Protocol) and the Mastodon / ActivityPub federation provide open, researcher-accessible data streams by design. Bluesky's "firehose" API provides real-time access to the entire public post stream at no cost. Mastodon's federated architecture allows researchers to access instance-level data with server administrator consent. These platforms are small relative to incumbents, but they are growing, and their architecture ensures that data access cannot be unilaterally revoked.

Government-mandated access. The EU's Digital Services Act (DSA), which took full effect in 2024, requires very large online platforms (VLOPs) to provide researchers with access to data for studying systemic risks. Article 40 establishes a framework for "vetted researcher" access, potentially creating a legal right to platform data for qualified academic researchers within the EU.

Synthetic data and simulation. LLM-based social simulations, where populations of AI agents interact under controlled conditions to model social dynamics, are emerging as a complement to observational research. These "silicon societies" allow researchers to test hypotheses about collective behavior at scale, without the ethical constraints of human experimentation and without dependency on platform APIs.

The transition is from a field dependent on corporate data access to one built on diverse, resilient data sources: open protocols, legally mandated access, donated personal data, controlled experiments, and computational simulation. Each source has limitations. The combination provides the methodological resilience that the Twitter-dependent era lacked.

Key Takeaway

Computational social science lost its primary data source when Twitter/X restricted API access (from free academic to ~$42K/month); Reddit followed. The field adapted through browser extension studies (algorithmic exposure tracking at NYU, Princeton), data donation protocols (GDPR-enabled), LLM-based digital twins (GPT-4 agents replicating survey responses at 85% accuracy), algorithmic auditing, and platform partnerships. The 2023 Meta/Science studies (n=23,000) found that algorithmic feeds increase ideological sorting, but short-term manipulation of individual features does not rapidly change political attitudes; the reshare mechanism is the primary amplification vector for misinformation. The fundamental challenge is the observer effect at scale: platform design shapes the behavior being measured, and no single platform is representative. The ethics asymmetry is stark: Facebook runs thousands of A/B tests daily without oversight, while academic researchers need IRB approval for 500 tweets. The EU's DSA Article 40 mandates researcher data access from very large platforms. The field is diversifying toward open protocols (Bluesky, Mastodon), synthetic simulation, and legally mandated access to reduce dependency on corporate goodwill.