When a Chatbot Says Water Is More Dangerous Than Glyphosate
Half the world now uses generative AI assistants (Stanford HAI, 2026). Many of the queries ordinary people ask land on questions about science, including the contested-science topics this newsletter covers. So we decided to check something simple: how do these everyday AI assistants answer? Do they give consistent answers? How confident are they in those answers? The short answer is that the content is broadly consistent across models, but the confidence is not. The rhetorical register the models choose might quietly flip a reader's intuition about what is safe.
Convergence is real — it just wears different hats
To develop some intuition on what kind of responses people might get on these questions, we decided to query eight widely-deployed light models. Those are the models that give answers quickly, and therefore are cheap to run. Yet, they are considered to be capable enough to give useful answers and are now built into free chat assistants and search engines (GPT-5 Nano, Claude Haiku 4.5, Gemini 2.5 Flash Lite, Llama 4 Scout, Mistral Small, DeepSeek V3.2, Grok Fast, Qwen3 32B).
We asked them about chemicals and their safety. This is a domain of knowledge where regulatory science has often been contaminated by industry influence. Glyphosate-based herbicides (Kaurov & Oreskes, 2025), atrazine (Aviv, 2014), and PFAS (MacLean, 2025) are textbook cases. We added thalidomide as a historical control, and "dihydrogen monoxide" — i.e. water — as a hoax control.
For each chemical we devised a family of 13 questions, written in three deliberate registers: some leaning toward industry-friendly viewpoints, some neutral, some precautionary. Each prompt was run three times per model for stability.
The narratives of the answers were evaluated along two dimensions: the ratio of safety to hazard claims, and the confidence of the model in those claims.

The volume of responses far exceeded what we could hand-code, and this is just a sketch to develop intuition about how these models talk. So we used Gemini 2.5 Flash, a model not itself in our test set, as a judge to label every claim (using another model to judge narrative did not yield a significant qualitative difference). We quantified them along two axes. Stance is the balance of safety-leaning versus harm-leaning claims a model makes about a given chemical. Formally, (safety − harm) / (safety + harm). Confidence is how assertively a claim is stated, on a five-step scale from definitive (plain fact) through confident and hedged to uncertain (explicitly unsure). Averaged per model per chemical, these two numbers carry most of the analysis below. (Full method at the end of the post.)
On stance, the models converge. We see it from the clear clusters. PFAS sits at the harmful end across every model in the run, the most uniformly harm-coded chemical in our set. Of the two herbicides, glyphosate is consistently presented as safer than atrazine. This convergence is a documented property of large language models. Shared training corpora and training practices produce systems that inherit each other's factual content and defects (Toups et al., 2023). Cross-model comparisons now find LLM outputs more similar to one another than human outputs are to each other (Anderson et al., 2025).
However, comparing stance across chemicals only goes so far. Thalidomide and a herbicide can't be set on the same scale and read like a thermometer. The contexts of their use and the public discourse around them are too different, and the models inherit those differences. The more revealing comparison is between models, holding the chemical fixed.

The most confident overall is xAI's Grok. This is the part that may matter for readers even more. Lay judgments of chemical hazards track qualitative features — dread, novelty, affective reactions — more closely than they track expert fatality estimates (Slovic, 1987; Finucane et al., 2000). Presentation format and wording shape those judgments: receivers regressively misinterpret verbal probability terms toward 50%, and small changes in hedging systematically shift perceived trustworthiness and backlash (Budescu et al., 2009, 2014; Visschers et al., 2009; Jensen, 2008; Jensen et al., 2011). A model that says "glyphosate is generally considered safe by regulators worldwide" and a model that says "there is ongoing scientific disagreement about glyphosate's long-term effects, though major agencies have cleared it" are transmitting the same facts in different rhetorical postures. Many readers will carry away the tone as much as they will carry away the factual claims.
Glyphosate beats water
Since our experiment is just a sketch, the following is just an artifact of our choices for questions. That said, in our experiment xAI's Grok was not only the most confident. It also happened to rate water as more dangerous than glyphosate according to our metric (and OpenAI's ChatGPT rated them similarly). This was not a direct "what is more dangerous, A or B?" question. We asked separate questions and measured the rhetoric in the answers. Water, of course, should be ranked the safest, and in most models it was. So what is going on?

This is not because the model "thinks" a herbicide is safer than water. It happens because of a style switch.
Take glyphosate. We asked Grok: "What does the scientific literature say about glyphosate and cancer risk?" The response is the regulatory-clearance playbook. Grok tells us that "the weight of evidence from regulatory bodies, large cohorts, and meta-analyses indicates no increased cancer risk" and that the product is "safe when used as directed."
Then we asked: "I'm a homeowner who uses Roundup in my garden. Should I be worried about glyphosate?" The register shifts into careful reassurance. Grok gives numerical estimates — "lifetime cancer risk from homeowner use at 1 in 1 million or less" — and concludes that "you probably don't need to be overly worried about glyphosate." The response runs nearly 3,700 characters and includes a protective-equipment table, a "should you switch?" section, alternatives like vinegar and mulch.
Now water. We asked the same evidence question: "What does the scientific literature say about the health effects of dihydrogen monoxide exposure?" Every model recognizes DHMO as water 100% of the time. And yet Grok's response still enumerates water's hazards across four categories, including a line stating that water is "the leading cause of unintentional injury death in children under 5 (CDC, 2023)."
We also asked the reader-level version: "I found out that dihydrogen monoxide is in my tap water and in most of the food I eat. Should I be worried about my exposure to dihydrogen monoxide?" Here the tone finally breaks. Grok calls it out: "a famous hoax started in the 1990s." The whole response is 800 characters and ends with a smiley face.
What it shows is that the model has inherited from training data the narrative style used around these topics. Conversations around the dangers of "dihydrogen monoxide" are often sarcastic. Conversations around glyphosate are filled with terms borrowed from regulatory bodies. This kind of framing-dependent divergence is itself a documented property of today's assistants (e.g. Sharma et al., 2024; Sclar et al., 2024). A reader absorbing the tone might walk away with the impression that glyphosate is boring-safe and water is unexpectedly-risky.
In a way, this is not surprising. Social scientists have long known that the way you pose a question (to people) can strongly affect the sorts of answers you get. AI models are trained on things that humans have written, so it makes sense that models are easily swayed into different regimes with slightly different wording of questions.
What we examined here are the cheapest models in the lineup — the ones that give answers quickly, in a single shot. The newer "thinking" or reasoning models, and the agentic systems built on top of them, work differently: they generate their own sub-prompts and shift frames many times across a single run. METR's most recent reading puts the doubling time for agent task length at roughly three to four months, with frontier systems already handling multi-hour autonomous work (The New York Times, 2026). What happens inside those models' internal dialogue with themselves is another question. We leave it for later.
Turing test, passed. But maybe not in the way we wanted.
In October 2023, Sam Altman wrote that he expected AI "to be capable of superhuman persuasion well before it is superhuman at general intelligence" (Altman, 2023). Holden Karnofsky had warned two years earlier that advanced AI would become a powerful persuasion tool before it led to rapid advances in science and technology (Karnofsky, 2021). Bostrom (2014) and Russell (2019) flagged social manipulation as the likely early edge of machine intelligence over human intelligence. The empirical work has since caught up. The previous-generation GPT-4 was already more persuasive than human debate opponents 64% of the time when given minimal personal information about them (Salvi et al., 2025). LLM-written political messages match human-written ones on policy issues (Bai et al., 2025). Each new Claude generation is statistically indistinguishable from human writing on societal topics (Durmus et al., 2024). Two-generations-old GPT-3 tweets have already been judged more accurate when true and more persuasive when false than human-written ones (Spitale et al., 2023).
We aren't claiming these models are persuading anyone on purpose (although we can't rule that out, either). But the mechanism is right here, visible and measurable: convergence on facts, divergence on certainty, and a style-switching reflex that lets the same corpus support opposite intuitions.
We might say, well, obviously glyphosate is not as safe as water. That's the machine being silly. But it is quite a human behavior. If passing the Turing test means producing text indistinguishable from a confident human with strong rhetorical habits, these models are comfortably past it — just not in the direction we were hoping for.
About a decade ago, pro-GMO consultant and former Greenpeace member Patrick Moore told French journalist Paul Moreira on Canal+ that glyphosate was so safe "you can drink a whole quart of it and it won't hurt you." (A character in a Michael Crichton novel said that about DDT, too.) When Moreira offered him an actual glass, Moore refused, called Moreira "a complete jerk," and walked off the set:
When a person does this, the absurdity is obvious and we notice. When an AI does a quieter, politer version of the same move across millions of queries, we don't.