• Dzso@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    9 hours ago

    I’m not saying these prompts won’t help, they probably will. But the notion that ChatGPT has any concept of “truth” is misleading. ChatGPT is a statistical language machine. It cannot evaluate truth. Period.

    • Zozano@aussie.zone
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      7 minutes ago

      What makes you think humans are better at evaluating truth? Most people can’t even define what they mean by “truth,” let alone apply epistemic rigor. Tweak it a little, and Gpt is more consistent and applies reasoning patterns that outperform the average human by miles.

      Epistemology isn’t some mystical art, it’s a structured method for assessing belief and justification, and large models approximate it surprisingly well. Sure it doesn’t “understand” truth in the human sense, but it does evaluate claims against internalized patterns of logic, evidence, and coherence based on a massive corpus of human discourse. That’s more than most people manage in a Facebook argument.

      So yes, it can evaluate truth. Not perfectly, but often better than the average person.