Demographic cues (eg, names, dialect) are widely used to study how LLM behavior may change depending on user demographics. Such cues are often assumed interchangeable.
π¨ We show they are not: different cues yield different model behavior for the same group and different conclusions on LLM bias. π§΅π
27.01.2026 13:07
π 18
π 10
π¬ 1
π 0
title: Cheap science, real harm: the cost of replacing human
participation with synthetic data
author: Abeba Birhane
abstract: Driven by the goals of augmenting diversity, increasing speed, reducing cost, the
use of synthetic data as a replacement for human participants is gaining traction
in AI research and product development. This talk critically examines the claim
that synthetic data can βaugment diversity,β arguing that this notion is empirically
unsubstantiated, conceptually flawed, and epistemically harmful. While speed and
cost-efficiency may be achievable, they often come at the expense of rigour, insight,
and robust science. Drawing on research from dataset audits, model evaluations,
Black feminist scholarship, and complexity science, I argue that replacing human
participants with synthetic data risks producing both real-world and epistemic
harms at worst and superficial knowledge and cheap science at best
I wrote this brief talk on why βaugmenting diversityβ with LLMs is empirically unsubstantiable, conceptually flawed, and epistemically harmful and a nice surprise to see the organisers have made it public
synthetic-data-workshop.github.io/papers/13.pdf
16.12.2025 10:57
π 827
π 260
π¬ 20
π 10
Auditing Google's AI Overviews and Featured Snippets: A Case Study on Baby Care and Pregnancy
Google Search increasingly surfaces AI-generated content through features like AI Overviews (AIO) and Featured Snippets (FS), which users frequently rely on despite having no control over their presen...
Google AI overviews now reach over 2B users worldwide. But how reliable are they on high stakes topics - for instance, pregnancy and baby care?
We have a new paper - led by Desheng Hu, now accepted at @icwsm.bsky.social - exploring that and finding many issues
Preprint: arxiv.org/abs/2511.12920
π§΅π
19.11.2025 16:58
π 16
π 9
π¬ 1
π 1
How public involvement can improve the science of AI | PNAS
As AI systems from decision-making algorithms to generative AI are deployed more widely,
computer scientists and social scientists alike are being ...
Can public involvement in AI evaluation improve the science? Or does it compromise quality, speed, cost?
In @pnas.org, Megan Price & I summarize challenges of AI evaluation, review strengths/weaknesses, & suggest how participatory methods can improve the science of AI
www.pnas.org/doi/10.1073/...
17.11.2025 12:47
π 19
π 13
π¬ 1
π 1
A large-scale field experiment on participatory decision-making in China - Nature Human Behaviour
Wu et al. show that involving citizens in local decision-making (participatory budgeting) improves civic engagement in a Chinese context.
Sherry Jueyu Wu showed that when people participate in collective decision-making, they are more willing to express that the gov needs improvement. Interesting to think about in the context of participation and accountability on online platforms...
π: www.nature.com/articles/s41...
18.11.2025 14:53
π 6
π 1
π¬ 1
π 0
had a lovely time at the News Futures workshop and my first CHI conf with some amazing folks πΈ
02.05.2025 00:45
π 3
π 0
π¬ 0
π 0
Large Language Models in Qualitative Research: Uses, Tensions, and Intentions | Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems
Excited to be presenting "LLMs in Qualitative Research: Uses, Tensions, and Intentions" with @mariannealq.bsky.social at #CHI2025 today!
π paper: dl.acm.org/doi/10.1145/...
27.04.2025 22:26
π 39
π 10
π¬ 2
π 0
I'm at #CHI2025! π―π΅ πΈ
Presenting our LBW "Traceable Texts and Their Effects".
We studied how phrase-level links from AI summaries to their sources influence the reading of complex texts.
π₯ April 30 at 10:30a & 3:40p πNorth 1F
Interested in text augmentation or improving source transparency? Drop by!
27.04.2025 01:44
π 10
π 3
π¬ 1
π 0
Whatβs Political on TikTok? A User-Centered Approach
Using audience perspectives and LLMs to map out the political landscape on TikTok.
LLMs show a lot of utility for analyzing content at scale. Here's @stephanietwang.bsky.social's write-up for GAIN on how she and collaborators used LLMs to examine political content on TikTok: generative-ai-newsroom.com/whats-politi...
25.04.2025 13:59
π 7
π 4
π¬ 0
π 0
Deeply evil
10.12.2024 04:01
π 2
π 0
π¬ 0
π 0
This @acm-cscw.bsky.social paper confirms the findings of the recent Nature paper on chrono feeds. Algorithmic feeds lead to more centrist, trustworthy content, but have little impact on user behavior. An independent audit of platforms w/o access to internal data!
dl.acm.org/doi/10.1145/...
25.11.2024 13:27
π 30
π 9
π¬ 0
π 0
Last week at CSCW, Catherine Han presented our work on journalists' unmet needs for protecting against harassment online. While the work targeted Twitter/X, it surfaces several nuances in users' needs that span future platforms as well (e.g., not wanting to filter out threats or visibly block users)
23.11.2024 18:38
π 30
π 9
π¬ 2
π 2
stopping by every coffee farm on the road is not helping my bean addiction βοΈ
14.11.2024 23:34
π 0
π 0
π¬ 0
π 0
hello world; participating in cscw 2024 + elections aftermath has finally convinced me to make a jump
14.11.2024 05:14
π 5
π 0
π¬ 0
π 0