Researchers create 30 fake student accounts to submit model-generated responses to real exams. Professors grade the 200 or 1500 word responses from the AI undergrads and gave them better grades than real students 84% of the time. 6% of the bot respondents did get caught, though… for being too good. Meanwhile, AI detection tools? Total bunk.

Will AI be the new calculator… or the death of us all (obviously the only alternative).

Note: the software was NOT as good on the advanced exams, even though it handled the easier stuff.

  • festus@lemmy.ca
    link
    fedilink
    English
    arrow-up
    77
    arrow-down
    3
    ·
    4 months ago

    Not at all surprising. ChatGPT ‘knows’ a course’s content insofar as it’s memorized the textbook and all the exam questions. Once you start asking it questions it’s never seen before (more likely for advanced topics that don’t have a billion study guides and tutorials for) it falls short, even for basic questions that’d just require a bit of additional logic.

    Mind you, memorizing everything is impressive and can get you a degree, but when tasked with a new problem never seen before ChatGPT is completely inadequate.

    • TheFriar@lemm.ee
      link
      fedilink
      English
      arrow-up
      26
      arrow-down
      1
      ·
      4 months ago

      Right? Can students use the internet on this test? Because the LLMs have the entire internet to search for the answers, and I guarantee you those textbooks and exam questions are online and searchable.

      • vortic@lemmy.world
        link
        fedilink
        English
        arrow-up
        18
        arrow-down
        1
        ·
        4 months ago

        I wonder how undergrads would do on the same exams given unlimited time and internet access but with LLMs blocked. That’s essentially what the LLMs have.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      4 months ago

      This is incorrect as was shown last year with the Skill-Mix research:

      Furthermore, simple probability calculations indicate that GPT-4’s reasonable performance on k=5 is suggestive of going beyond “stochastic parrot” behavior (Bender et al., 2021), i.e., it combines skills in ways that it had not seen during training.