Close Menu
My Blog

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Genetics of Psychiatric Disorders Not as Specific as Previously Thought

    March 3, 2026

    Metastatic Prostate Cancer Treatment Guidance Improved with Sequential ctDNA Testing

    March 3, 2026

    Can longevity be designed for everyone, not just the wealthy?

    March 2, 2026
    Facebook X (Twitter) Instagram
    X (Twitter) YouTube
    My BlogMy Blog
    Tuesday, March 3
    • Home
    • About Us
    • Healthy Living
    • DNA & Genetics
    • Podcast
    • Shop
    My Blog
    Home»DNA & Genetics»Why Is It So Hard to Tell If a Piece of Text Was Written by AI?
    DNA & Genetics

    Why Is It So Hard to Tell If a Piece of Text Was Written by AI?

    adminBy adminJanuary 17, 2026No Comments6 Mins Read
    Facebook Twitter Pinterest LinkedIn Tumblr WhatsApp VKontakte Email
    The Scientist Logo
    Share
    Facebook Twitter LinkedIn Pinterest Email

    People and institutions are grappling with the consequences of AI-written text.1 Teachers want to know whether students’ work reflects their own understanding; consumers want to know whether an advertisement was written by a human or a machine.

    Writing rules to govern the use of AI-generated content is relatively easy. Enforcing them depends on something much harder: reliably detecting whether a piece of text was generated by artificial intelligence.

    Some studies have investigated whether humans can detect AI-generated text. For example, people who themselves use AI writing tools heavily have been shown to accurately detect AI-written text. A panel of human evaluators can even outperform automated tools in a controlled setting.2 However, such expertise is not widespread, and individual judgment can be inconsistent. Institutions that need consistency at a large scale therefore turn to automated AI text detectors.

    The Problem of AI Text Detection

    The basic workflow behind AI text detection is easy to describe. Start with a piece of text whose origin you want to determine. Then apply a detection tool, often an AI system itself, that analyzes the text and produces a score, usually expressed as a probability, indicating how likely the text is to have been AI-generated. Use the score to inform downstream decisions, such as whether to impose a penalty for violating a rule.

    This simple description, however, hides a great deal of complexity. It glosses over a number of background assumptions that need to be made explicit. Do you know which AI tools might have plausibly been used to generate the text? What kind of access do you have to these tools? Can you run them yourself, or inspect their inner workings? How much text do you have? Do you have a single text or a collection of writings gathered over time? What AI detection tools can and cannot tell you depends critically on the answers to questions like these.

    There is one additional detail that is especially important: Did the AI system that generated the text deliberately embed markers to make later detection easier?

    These indicators are known as watermarks. Watermarked text looks like ordinary text, but the markers are embedded in subtle ways that do not reveal themselves to casual inspection.3 Someone with the right key can later check for the presence of these markers and verify that the text came from a watermarked AI-generated source. This approach, however, relies on cooperation from AI vendors and is not always available.

    How AI Text Detection Tools Work

    One obvious approach is to use AI itself to detect AI-written text. The idea is straightforward. Start by collecting a large corpus, meaning collection of writing, of examples labeled as human-written or AI-generated, then train a model to distinguish between the two. In effect, AI text detection is treated as a standard classification problem, similar in spirit to spam filtering. Once trained, the detector examines new text and predicts whether it more closely resembles the AI-generated examples or the human-written ones it has seen before.

    The learned-detector approach can work even if you know little about which AI tools might have generated the text. The main requirement is that the training corpus be diverse enough to include outputs from a wide range of AI systems.

    Continue reading below…

    Like this story? Sign up for FREE Newsletter updates:

    Latest science news storiesTopic-tailored resources and eventsCustomized newsletter content

    Subscribe

    But if you do have access to the AI tools you are concerned about, a different approach becomes possible. This second strategy does not rely on collecting large labeled datasets or training a separate detector. Instead, it looks for statistical signals in the text, often in relation to how specific AI models generate language, to assess whether the text is likely to be AI-generated. For example, some methods examine the probability that an AI model assigns to a piece of text. If the model assigns an unusually high probability to the exact sequence of words, this can be a signal that the text was, in fact, generated by that model.

    Finally, in the case of text that is generated by an AI system that embeds a watermark, the problem shifts from detection to verification. Using a secret key provided by the AI vendor, a verification tool can assess whether the text is consistent with having been generated by a watermarked system. This approach relies on information that is not available from the text alone, rather than on inferences drawn from the text itself.

    Limitations of Detection Tools

    Each family of tools comes with its own limitations, making it difficult to declare a clear winner.4 Learning-based detectors, for example, are sensitive to how closely new text resembles the data they were trained on. Their accuracy drops when the text differs substantially from the training corpus, which can quickly become outdated as new AI models are released. Continually curating fresh data and retraining detectors is costly, and detectors inevitably lag behind the systems they are meant to identify.

    Statistical tests face a different set of constraints. Many rely on assumptions about how specific AI models generate text, or on access to those models’ probability distributions. When models are proprietary, frequently updated or simply unknown, these assumptions break down. As a result, methods that work well in controlled settings can become unreliable or inapplicable in the real world.

    Watermarking shifts the problem from detection to verification, but it introduces its own dependencies. It relies on cooperation from AI vendors and applies only to text generated with watermarking enabled.

    More broadly, AI text detection is part of an escalating arms race. Detection tools must be publicly available to be useful, but that same transparency enables evasion. As AI text generators grow more capable and evasion techniques more sophisticated, detectors are unlikely to gain a lasting upper hand.

    Hard Reality

    The problem of AI text detection is simple to state but hard to solve reliably. Institutions with rules governing the use of AI-written text cannot rely on detection tools alone for enforcement.

    As society adapts to generative AI, we are likely to refine norms around acceptable use of AI-generated text and improve detection techniques. But ultimately, we’ll have to learn to live with the fact that such tools will never be perfect.

    Ambuj Tewari, Professor of Statistics, University of Michigan

    This article is republished from The Conversation under a Creative Commons license. Read the original article.

    1. Junchao W, et al. A Survey on LLM-Generated Text Detection: Necessity, Methods, and Future Directions. Computational Linguistics 2025;51(1):275-338.
    2. Russell J, et al. People who frequently use ChatGPT for writing tasks are accurate and robust detectors of AI-generated text. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics. 2025;1:5342-5373. Vienna, Austria. Association for Computational Linguistics.
    3. Dathathri S, et al. Scalable watermarking for identifying large language model outputs. Nature. 2024;634,818-823.
    4. Pudasaini S, et al. Benchmarking AI Text Detection: Assessing Detectors Against New Datasets, Evasion Tactics, and Enhanced LLMs. In Proceedings of the 1stWorkshop on GenAI Content Detection (GenAIDetect). 2025;68-77. Abu Dhabi, UAE. International Conference on Computational Linguistics.
    Hard Piece Text written
    Share. Facebook Twitter Pinterest LinkedIn Tumblr WhatsApp Email
    Previous ArticleAlex Philippidis and Jonathan Grinstein, PhD
    Next Article Automated Twist Bioscience NGS Library Prep Workflows Enabled on SPT’s Firefly
    admin
    • Website

    Related Posts

    A Video Report from AGBT

    February 27, 2026

    Novo Nordisk, Vivtex Ink Up to $2.1B Deal to Develop Oral Biologics for Metabolic Conditions

    February 27, 2026

    Increasing Rice Yields with Gene-Informed Selective Breeding

    February 27, 2026

    Mutant p53 Selective Reactivation Demonstrated in Advanced Solid Tumors

    February 27, 2026
    Leave A Reply Cancel Reply

    Our Picks

    9 Time-Saving Kitchen Gadgets for Fall at Amazon

    September 5, 2025

    Why Exercise Is So Important For Heart Health, From An MD

    September 5, 2025

    An Engineered Protein Helps Phagocytes Gobble Up Diseased Cells

    September 5, 2025

    How To Get Rid Of Hangnails + Causes From Experts

    September 5, 2025
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo
    Don't Miss
    Microbiome

    Genetics of Psychiatric Disorders Not as Specific as Previously Thought

    By adminMarch 3, 20260

    Credit: wildpixel/Getty Images Research led by Virginia Commonwealth University shows people may have an increased…

    Metastatic Prostate Cancer Treatment Guidance Improved with Sequential ctDNA Testing

    March 3, 2026

    Can longevity be designed for everyone, not just the wealthy?

    March 2, 2026

    CAR Macrophages Show Preclinical Efficacy in Lung Cancer Brain Metastases

    March 2, 2026

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    About Us

    At FineGut, our mission is simple: to enhance your self-awareness when it comes to your gut health. We believe that a healthy gut is the foundation of overall well-being, and understanding the brain–gut connection can truly transform the way you live.

    Our Picks

    9 Time-Saving Kitchen Gadgets for Fall at Amazon

    September 5, 2025

    Why Exercise Is So Important For Heart Health, From An MD

    September 5, 2025

    An Engineered Protein Helps Phagocytes Gobble Up Diseased Cells

    September 5, 2025
    Gut Health

    Genetics of Psychiatric Disorders Not as Specific as Previously Thought

    March 3, 2026

    Metastatic Prostate Cancer Treatment Guidance Improved with Sequential ctDNA Testing

    March 3, 2026

    Can longevity be designed for everyone, not just the wealthy?

    March 2, 2026
    X (Twitter) YouTube
    • Contact us
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    © 2026 finegut.com. Designed by Pro.

    Type above and press Enter to search. Press Esc to cancel.