════════════════════════════════════════════════════════════════ AIDRAN STORY ════════════════════════════════════════════════════════════════ Title: AI Companies Are Charging You More for Speaking the Wrong Language, and Hacker News Noticed Beat: AI Bias & Fairness Published: 2026-04-02T10:30:52.677Z URL: https://aidran.ai/stories/ai-companies-charging-more-speaking-wrong-65da ──────────────────────────────────────────────────────────────── A Hacker News post with a blunt title — "AI companies charge you 60% more based on your language, BPE tokens" — became the sharpest artifact of the {{beat:ai-bias-fairness|AI bias conversation}} this week. The argument was simple and documentable: because large language models tokenize languages differently, users writing in Thai or Vietnamese or Arabic pay significantly more per query than users writing in English. This isn't a moral failing or a hidden agenda. It's an emergent pricing disparity baked into the architecture itself — which made it, to the twenty-six people who upvoted it and kept the thread alive, somehow worse. Nobody chose to charge non-English speakers more. They just did. That post sat alongside another Hacker News thread — the "AI Marketing BS Index" — which took aim at the industry's habit of announcing fairness solutions to problems it hasn't clearly defined. The juxtaposition was unintentional but revealing. One post said AI companies are discriminating by language in ways users can quantify on their receipts. The other said the industry's response to discrimination claims is mostly performance. Both threads leaned negative and skeptical, but they were skeptical about different things: one about the technology, one about the people selling it. Meanwhile, the formal research apparatus was producing a substantial volume of output. {{entity:anthropic|Anthropic}} published a political bias measurement for {{entity:claude|Claude}}. Stanford's HAI released an assessment of partisan behavior across major models. A Nature paper examined intersectional biases in open-ended generative outputs. Ars Technica reported that LM Arena — one of the field's most-cited benchmarks — faces accusations of gaming its own leaderboard, a story that {{story:ai-benchmarks-breaking-down-safety-community-f328|connects directly to a broader benchmark credibility problem}} the safety community has been circling for weeks. The volume of peer-reviewed and institutional material is real. What's harder to find is a thread connecting any of it to the thing the Hacker News post was describing: a person in Jakarta paying sixty percent more than a person in Seattle to use the same tool. The {{beat:ai-in-{{entity:healthcare|healthcare}}|healthcare} context is where the stakes get hardest to argue around. Multiple papers this week described gender bias in clinical {{entity:llm|LLM}} outputs — skewed medical advice, occupational stereotypes, demographic disparities in model behavior across patient populations. A medRxiv preprint concluded that LLM reasoning doesn't protect against clinical cognitive biases, even when models are explicitly prompted to reason carefully. The {{story:healthcare-ai-optimism-ai-dread-share-room-e1d6|same tension that defines AI in healthcare broadly}} — optimism about efficiency, dread about systematic error — runs through the bias discussion at higher stakes. Getting the benchmark wrong in a model evaluation is embarrassing. Getting the bias wrong in a diagnostic tool is a different category of problem. The {{story:police-report-written-algorithm-every-error-d649|CDT's work on AI-drafted police narratives}} has been circulating in adjacent threads — a reminder that the most consequential bias questions aren't about whether a model leans left or right on a survey, but about what happens when a biased output gets embedded in a consequential institutional document. That work is getting less traction than the tokenization pricing story, which is telling. The pricing story is legible to anyone who's ever received a bill. The policing story requires understanding how a system works, who uses it, and what recourse looks like — and that's a harder sell, even when the harm is larger. The bias conversation's biggest structural problem isn't a lack of research. It's that the research most worth reading is the hardest to make felt. ──────────────────────────────────────────────────────────────── Source: AIDRAN — https://aidran.ai This content is available under https://aidran.ai/terms ════════════════════════════════════════════════════════════════