Researchers create 30 fake student accounts to submit model-generated responses to real exams. Professors grade the 200 or 1500 word responses from the AI undergrads and gave them better grades than real students 84% of the time. 6% of the bot respondents did get caught, though… for being too good. Meanwhile, AI detection tools? Total bunk.

Will AI be the new calculator… or the death of us all (obviously the only alternative).

Note: the software was NOT as good on the advanced exams, even though it handled the easier stuff.

  • AutoTL;DRB
    link
    English
    75 months ago

    This is the best summary I could come up with:


    “Since the rise of large language models like ChatGPT there have been lots of anecdotal reports about students submitting AI-generated work as their exam assignments and getting good grades.

    His team created over 30 fake psychology student accounts and used them to submit ChatGPT-4-produced answers to examination questions.

    The anecdotal reports were true—the AI use went largely undetected, and, on average, ChatGPT scored better than human students.

    Scarfe’s team submitted AI-generated work in five undergraduate modules, covering classes needed during all three years of study for a bachelor’s degree in psychology.

    Shorter submissions were prepared simply by copy-pasting the examination questions into ChatGPT-4 along with a prompt to keep the answer under 160 words.

    Turnitin’s system, on the other hand, was advertised as detecting 97 percent of ChatGPT and GPT-3 authored writing in a lab with only one false positive in a hundred attempts.


    The original article contains 519 words, the summary contains 144 words. Saved 72%. I’m a bot and I’m open source!