Does AI actually help students learn? A recent experiment in a high school provides a cautionary tale.

Researchers at the University of Pennsylvania found that Turkish high school students who had access to ChatGPT while doing practice math problems did worse on a math test compared with students who didn’t have access to ChatGPT. Those with ChatGPT solved 48 percent more of the practice problems correctly, but they ultimately scored 17 percent worse on a test of the topic that the students were learning.

A third group of students had access to a revised version of ChatGPT that functioned more like a tutor. This chatbot was programmed to provide hints without directly divulging the answer. The students who used it did spectacularly better on the practice problems, solving 127 percent more of them correctly compared with students who did their practice work without any high-tech aids. But on a test afterwards, these AI-tutored students did no better. Students who just did their practice problems the old fashioned way — on their own — matched their test scores.

  • MystikIncarnate@lemmy.ca
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Something I’ve noticed with institutional education is that they’re not looking for the factually correct answer, they’re looking for the answer that matches whatever you were told in class. Those two things should not be different, but in my experience, they’re not always the same thing.

    I have no idea if this is a factor here, but it’s something I’ve noticed. I have actually answered questions with a factually wrong answer, because that’s what was taught, just to get the marks.

  • 2ugly2live@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    I don’t even know of this is ChatGPT’s fault. This would be the same outcome if someone just gave them the answers to a study packet. Yes, they’ll have the answers because someone (or something) gave it to them, but won’t know how to get that answer without teaching them. Surprise: For kids to learn, they need to be taught. Shocker.

    • Buddahriffic@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      I’ve found chatGPT to be a great learning aid. You just don’t use it to jump straight to the answers, you use it to explore the gaps and edges of what you know or understand. Add context and details, not final answers.

      • IzzyScissor@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        The study shows that once you remove the LLM though, the benefit disappears. If you rely on an LLM to help break things down or add context and details, you don’t learn those skills on your own.

        I used it to learn some coding, but without using it again, I couldn’t replicate my own code. It’s a struggle, but I don’t think using it as a teaching aid is a good idea yet, maybe ever.

        • jpeps@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          I wouldn’t say this matches my experience. I’ve used LLMs to improve my understanding of a topic I’m already skilled in, and I’m just looking to understand something nuanced. Being able to interrogate on a very specific question that I can appreciate the answer to is really useful and definitely sticks with me beyond the chat.

        • obbeel@lemmy.eco.br
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          There are lots of studies out there, and many of them contradict each other. Having a study with references contribute to the discussion, but it isn’t the absolute truth.

  • N0body@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Traditional instruction gave the same result as a bleeding edge ChatGPT tutorial bot. Imagine what would happen if a tiny fraction of the billions spent to develop this technology went into funding improved traditional instruction.

    Better paid teachers, better resources, studies geared at optimizing traditional instruction, etc.

    Move fast and break things was always a stupid goal. Turbocharging it with all this money is killing the tried and true options that actually produce results, while straining the power grid and worsening global warming.

    • technocrit@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      The education system is primarily about controlling bodies and minds. So any actual education is counter-productive.

      • elvith@feddit.org
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        It’s the other way round: Education makes for less gullible people and for workers that demand more rights more freely and easily - and then those are coming for their yachts…

    • otp@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Traditional instruction gave the same result as a bleeding edge ChatGPT tutorial bot.

      Interesting way of looking at it. I disagree with your conclusion about the study, though.

      It seems like the AI tool would be helpful for things like assignments rather than tests. I think it’s intellectually dishonest to ignore the gains in some environments because it doesn’t have gains in others.

      You’re also comparing a young technology to methods that have been adapted over hundreds of thousands of years. Was the first automobile entirely superior to every horse?

      I get that some people just hate AI because it’s AI. For the people interested in nuance, I think this study is interesting. I think other studies will seek to build on it.

      • Kalysta@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        The point of assignments is to help study for your test.

        Homework is forced study. If you’re just handed the answers, you will do shit on the test.

        • otp@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          The point of assignments is to help study for your test.

          To me, “assignment” is more of a project. Not rote practice. Applying knowledge to a bit of a longer term, multi-part project.

    • Petter1@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Imagine all the money spent on war would be invested into education 🫣what a beautiful world we would live in.

    • littlewonder@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      The first sentence of this comment says everything. If a technology that is still ironing out its capabilities is able to get kids almost to the level of in-person instruction, think of the potential when used in tandem with teachers, or even when it has matured into a polished version of itself.

      How many of these kids knew how to leverage a GPT while avoiding common pitfalls? Would they have performed even better if given info on creating prompts for studying?

      LLMs/GPT, and other forms of the AI boogeyman, are all just a tool we can use to augment education when it makes sense. Just like the introduction of calculators or the internet, AI isn’t going to be the easy button, nor is it going to steal all teacher’s jobs. These tools need to be studied, trained for, and applied purposely in order to be most effective.

      • petrol_sniff_king@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        are all just a tool
        just a tool
        it’s just a tool
        a tool is a tool
        all are just tools
        it’s no more than a tool
        it’s just a tool
        it’s a tool we can use
        one of our many tools
        it’s only a tool
        these are just tools
        a tool for thee, a tool for me

        guns don’t kill people, people kill people
        the solution is simple:
        teach drunk people not to shoot their guns so much
        unless they want to
        that is the American way

        tanks don’t kill people, people kill people
        the solution is simple:
        teach drunk people not to shoot their tanks so much
        the barista who offered them soy milk
        wasn’t implying anything about their T levels
        that is the American way

        Thanks for reminding me that AI is just tools, friend.
        My memory is not so good.
        I often can’t
        remember

        • littlewonder@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          2 months ago

          Ok, I’m going to reply like you’re being serious. It is a tool and it’s out there and it’s not going anywhere. Do we allow ourselves to imagine how it can be improved to help students or do we ignore it and act like it won’t ever be something students need to learn?

  • glowie@h4x0r.host
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Of all the students in the world, they pick ones from a “Turkish high school”. Any clear indication why there of all places when conducted by a US university?

    • Phoenix3875@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      The paper only says it’s a collaboration. It’s pretty large scale, so the opportunity might be rare. There’s a chance that (the same or other) researchers will follow up and experiment in more schools.

    • Lemminary@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      If I had access to ChatGPT during my college years and it helped me parse things I didn’t fully understand from the texts or provided much-needed context for what I was studying, I would’ve done much better having integrated my learning. That’s one of the areas where ChatGPT shines. I only got there on my way out. But math problems? Ugh.

      • ForgotAboutDre@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        When you automate these processes you lose the experience. I wouldn’t be surprised if you couldn’t parse information as well as you can now, if you had access to chat GPT.

        It’s had to get better at solving your problems if something else does it for you.

        Also the reliability of these systems is poor, and they’re specifically trained to produce output that appears correct. Not actually is correct.

        • Veddit@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          I read that comment, and use it similarly, as more a super-dictionary/encyclopedia in the same way I’d watch supplementary YouTube videos to enhance my understanding. Rather than automating the understanding process.

          More like having a tutor who you ask all the too-stupid and too-hard questions to, who never gets tired or fed up with you.

          • Petter1@lemm.ee
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 months ago

            Exactly this! That is why I always have at least one instance of AI chatbot running when I am coding or better said analyse code for debugging.

            It makes it possible to debug kernel stuff without much pre-knowledge, if you are proficient in prompting your questions. Well, it did work for me.

        • Lemminary@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          I quickly learned how ChatGPT works so I’m aware of its limitations. And since I’m talking about university students, I’m fairly sure those smart cookies can figure it out themselves. The thing is, studying the biological sciences requires you to understand other subjects you haven’t learned yet, and having someone explain how that fits into the overall picture puts you way ahead of the curve because you start integrating knowledge earlier. You only get that from retrospection once you’ve passed all your classes and have a panoramic view of the field, which, in my opinion, is too late for excellent grades. This is why I think having parents with degrees in a related field or personal tutors gives an incredibly unfair advantage to anyone in college. That’s what ChatGPT gives you for free. Your parents and the tutors will also make mistakes, but that doesn’t take away the value which is also true for the AIs.

          And regarding the output that appears correct, some tools help mitigate that. I’ve used the Consensus plugin to some degree and think it’s fairly accurate for resolving some questions based on research. What’s more useful is that it’ll cite the paper directly so you can learn more instead of relying on ChatGPT alone. It’s a great tool I wish I had that would’ve saved me so much time to focus on other more important things instead of going down the list of fruitless search results with a million tabs open.

          One thing I will agree with you is probably learning how to use Google Scholar and Google Books and pirating books using the library to find the exact information as it appears in the textbooks to answer homework questions which I did meticulously down to the paragraph. But only I did that. Everybody else copied their homework, so at least in my university it was a personal choice how far you wanted to take those skills. So now instead of your peers giving you the answers, it’s ChatGPT. So my question is, are we really losing anything?

          Overall I think other skills need honing today, particularly verifying information, together with critical thinking which is always relevant. And the former is only hard because it’s tedious work, honestly.

    • catloaf@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      I’m guessing there was a previous connection with some of the study authors.

      I skimmed the paper, and I didn’t see it mention language. I’d be more interested to know if they were using ChatGPT in English or Turkish, and how that would affect performance, since I assume the model is trained on significantly more English language data than Turkish.

      • Em Adespoton@lemmy.ca
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        GPTs are designed with translation in mind, so I could see it being extremely useful in providing me instruction on a topic in a non-English native language.

        But they haven’t been around long enough for the novelty factor to wear off.

        It’s like computers in the 1980s… people played Oregon Trail on them, but they didn’t really help much with general education.

        Fast forward to today, and computers are the core of many facets of education, allowing students to learn knowledge and skills that they’d otherwise have no access to.

        GPTs will eventually go the same way.

    • technocrit@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      2 months ago

      The study was done in Turkey, probably because students are for sale and have no rights.

      It doesn’t matter though. They could pick any weird, tiny sample and do another meaningless study. It would still get hyped and they would still get funding.

  • vin@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Did those using tutor AI spend less time on learning? That would have been worth measuring

  • trougnouf@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    The title is pretty misleading. Kids who used ChatGPT to get hints/explanations rather than outright getting the answers did as well as those who had no access to ChatGPT. They probably had a much easier time studying/understanding with it so it’s a win for LLMs as a teaching tool imo.

    • Apollo42@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      2 months ago

      Is it really a win for LLMs if the study found no significant difference between those using it as a tutor and those not?

      • 8andage@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 months ago

        Maybe using llm assistance was less stressful or quicker than self study. The tutoring focused llm is definitely better than allowing full access to gpt itself, which is what is currently happening

      • EatATaco@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        As another poster questioned, if it saved them tine then, yes, it is absolutely a win. But if they spent the same amount of time, I would agree with you that it’s not a win.

      • trougnouf@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Not everyone can afford a tutor or knows where to find an expert that can answer questions in any given domain. I think such a tool would have made understanding a lot of my college courses a lot easier.

  • flerp@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Like any tool, it depends how you use it. I have been learning a lot of math recently and have been chatting with AI to increase my understanding of the concepts. There are times when the textbook shows some steps that I don’t understand why they’re happening and I’ve questioned AI about it. Sometimes it takes a few tries of asking until you figure out the right question to ask to get the right answer you need, but that process of thinking helps you along the way anyways by crystallizing in your brain what exactly it is that you don’t understand.

    I have found it to be a very helpful tool in my educational path. However I am learning things because I want to understand them, not because I have to pass a test and that determination in me to want to understand is a big difference. Just getting hints to help you solve the problem might not really help in the long run, but it you’re actually curious about what you’re learning and focus on getting a deeper understanding of why and how something works rather than just getting the right answer, it can be a very useful tool.

    • Gsus4@mander.xyz
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      2 months ago

      Sometimes it leads me wildly astray when I do that, like a really bad tutor…but it is good if you want a refresher and can spot the bullshit on the side. It is good for spotting things that you didnt know before and can factcheck afterwards.

      …but maybe other review papers and textbooks are still better…

    • Rekorse@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Why are you so confident that the things you are learning from AI are correct? Are you just using it to gather other sources to review by hand or are you trying to have conversations with the AI?

      We’ve all seen AI get the correct answer but the show your work part is nonsense, or vice versa. How do you verify what AI outputs to you?

      • Buttons@programming.dev
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 months ago

        I, like the OP, was also studying math from a textbook and using GPT4 to help clear things up. GPT4 caught an error in the textbook.

        The LLM doesn’t have a theory of mind, it wont start over and try to explain a concept from a completely new angle, it mostly just repeats the same stuff over and over. Still, once I have figured something out, I can ask the LLM if my ideas are correct and it sometimes makes small corrections.

        Overall, most of my learning came from the textbook, and talking with the LLM about the concepts I had learned helped cement them in my brain. I didn’t learn a whole lot from the LLM directly, but it was good enough to confirm what I learned from the textbook and sometimes correct mistakes.

      • GaMEChld@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        You check it’s work. I used it to calculate efficiency in a factory game and went through and made corrections to inconsistencies I spotted. Always check it’s work.

        • flerp@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          Exactly. It’s a helpful tool but it needs to be used responsibly. Writing it off completely is as bad a take as blindly accepting everything it spits out.

      • pflanzenregal@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        I use it for explaining stuff when studying for uni and I do it like this: If I don’t understand e.g. a definition, I ask an LLM to explain it, read the original definition again and see if it makes sense.

        This is an informal approach, but if the definition is sufficiently complex, false answers are unlikely to lead to an understanding. Not impossible ofc, so always be wary.

        For context: I’m studying computer science, so lots of math and theoretical computer science.

      • flerp@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        I’m not at all confident in the answers directly. I’ve gotten plenty of wrong answers form AI and I’ve gotten plenty of correct answers. If anything it’s just more practice for critical thinking skills, separating what is true and what isn’t.

        When it comes to math though, it’s pretty straightforward, I’m just looking for context on some steps in the problems, maybe reminders of things I learned years ago and have forgotten, that sort of thing. As I said, I’m interested in actually understanding the stuff that I’m learning because I am using it for the things I’m working on so I’m mainly reading through textbooks and using AI as well as other sources online to round out my understanding of the concepts. If I’m getting the right answers and the things I am doing are working, it’s a good indicator I’m on the right path.

        It’s not like I’m doing cutting edge physics or medical research where mistakes could cause lives.

        • Rekorse@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          Its sort of similar to saying poppy production overall is pretty negative, but if smart critical people use it sparingly and apprehensively, opiates could be of great benefit to that person.

          Thats all well and good and all but AI is not being developed to help critical thinkers research slightly easier, its being created to reduce the amount of money companies spend on humans.

          Until regulations are in place to guide the development of the technology in useful ways then I dont know any of it should be permitted. What’s the rush for anyways?

          • flerp@lemm.ee
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 months ago

            Well I’m definitely not pushing for more AI and I like to try to stay nuanced on the topic. Like I mentioned in my first comment I have found it to be a very helpful tool but if used in other ways it could do more harm than good. I’m not involved in making or pushing AI but as long as it is an available tool I’m going to make use of it in the most responsible way I can and talk about how I use it knowing that I can’t control what other people do but maybe I could help some people who are only using it to get answer hints like in the article to find more useful ways of using it.

            When it comes to regulation, yeah I’m all for that. It’s a sad reality that regulation always lags behind and generally doesn’t get implemented until there’s some sort of problem that scares the people in power who are mostly too old to understand what’s happening anyways.

            And as to what’s the rush, I would say a combination of curiosity and good intentions mixed with the worst of capitalism, the carrot of financial gain for success and the stick of financial ruin for failure and I don’t have a clue what percent of the pie each part makes up. I’m not saying it’s a good situation but it’s the way things go and I don’t think anyone alive could stop it. Once something is out of the bag, there ain’t any putting it back.

            Basically I’m with you that it will be used for things that make life worse for people and that sucks, and it would be great if that was not the case but that doesn’t change the fact that I can’t do anything about that and meanwhile it can still be a useful tool and so I’m going to use it the best that I can regardless how others use it because there’s really nothing I can do except keep pushing forward the best I can, just like anyone else.

      • NιƙƙιDιɱҽʂ@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 months ago

        I personally use it’s answers as a jumping off point to do my own research, or I ask it for sources directly about things and check those out. I frequently use LLMs for learning about topics, but definitely don’t take anything they say at face value.

        For a personal example, I use ChatGPT as my personal Japanese tutor. I use it discuss and break down nuances of various words or sayings, names of certain conjugation forms etc. etc., and it is absolutely not 100% correct, but I can now take the names of things that it gives me in native Japanese that I never would have known and look them up using other resources. Either it’s correct and I find confirming information, or it’s wrong and I can research further independently or ask it follow up questions. It’s certainly not as good as a human native speaker, but for $20 a month and as someone who likes enjoys doing their own research, I fucking love it.

        • obbeel@lemmy.eco.br
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          Hey, that’s a cool thing to do! I’ll try it. Learning a new language through LLMs sounds cool.

      • KairuByte@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        I mean, why are you confident the work in textbooks is correct? Both have been proven unreliable, though I will admit LLMs are much more so.

        The way you verify in this instance is actually going through the work yourself after you’ve been shown sources. They are explicitly not saying they take 1+1=3 as law, but instead asking how that was reached and working off that explanation to see if it makes sense and learn more.

        Math is likely the best for this too. You have undeniable truths in math, it’s true, or it’s false. There are no (meaningful) opinions on how addition works other than the correct one.

        • Rekorse@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          The problem with this style of verification is that there is no authoritative source. Neither the AI nor yourself is capable of verifying for accuracy. The AI also has no expectation of being accurate or revised.

          I don’t see how this is any better than running google searches on reddit or other message boards looking for relevant discussions and basing your knowledge on those.

          If AI was enabling something new that might be worth it but allowing someone to find slightly less/more shitty message board posts 10% more efficiently isnt worth what’s happening. There are countries that are capable of regulation as a field fills out, why can’t america? We banned tiktok in under a month didnt we?

  • Allero@lemmy.today
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Perhaps unsurprisingly. Any sort of “assistance” with answers will do that.

    Students have to learn why things work the way they do, and they won’t be able to grasp it without going ahead and doing every piece manually.

  • ???@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Yeh because it’s just like having their dumb parents do homework for them

  • prosp3kt@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    2 months ago

    There are a part here that sounds interesting

    The students who used it did spectacularly better on the practice problems, solving 127 percent more of them correctly compared with students who did their practice work without any high-tech aids. But on a test afterwards, these AI-tutored students did no better.

    Do you think that these students that used ChatGPT can do the exercises “the old fashioned way”? For me it was a nightmare try to resolve a calculus problem just with the trash books that doesn’t explain a damn fuck, I have to go to different resources, wolphram, youtube, but what happened when there was a problem that wasnt well explained in any resource?. I hate openAI, I want to punch Altman in the face. But this doesn’t mean we have to bait this hard in the title.

    • MBM@lemmings.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Maybe it’s an accident but you left out your fellow students and the teacher, in my eyes the most useful resources

  • Maggoty@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    ChatGPT lies which is kind of an issue in education.

    As far as seeing the answer, I learned a significant amount of math by looking at the answer for a type of question and working backwards. That’s not the issue as long as you’re honestly trying to understand the process.

  • PriorityMotif@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    There’s a bunch of websites that give you the answers to most homework. You can just Google the question and find the answers pretty quickly. I assume the people using chatgtp to “study” are just cheating on homework anyway.