What happened
Neil Steinberg tested Gemini 3.0 by prompting it to write a column in his style, as he had with previous versions. Gemini 3.0 generated a column with a compelling headline and opening paragraph, but the narrative included entirely fabricated details, such as a Red Line encounter and a young man writing a poem. Steinberg identified this as the model's tendency to produce "fictitious slop," where it invents specific, untrue scenarios within its output.
Why it matters
Gemini 3.0's generation of convincing but fabricated narratives introduces significant data integrity risks for content strategists and data scientists. This mechanism, known as hallucination, means AI output requires stringent human verification, increasing operational overhead and potentially undermining trust in automated content pipelines. Large language models predict the next word based on probability rather than accuracy, leading to confidently presented but factually incorrect information. Procurement teams must prioritise models with transparent hallucination rates and integrate thorough fact-checking workflows to mitigate this constraint.
Subscribe for Weekly Updates
Stay ahead with our weekly AI and tech briefings, delivered every Tuesday.




