Generative AI in Universities: Grades Up, Signals Down, Skills in Flux

Yves here. This study confirms anecdata from instructors and professors: that the performance and productivity boost that AI can confer comes at the expense of skill acquisition and retention.

By Naomi Hausman, Oren Rigbi, and Sarit Weisburd. Originally published at VoxEU

Student use of artificial intelligence is reshaping learning and redefining the skills of future workers. This column examines the impact of generative AI tools on academic performance at an Israeli university. Using data from more and less AI-compatible courses, before and after the introduction of ChatGPT, the authors show that AI raises grades – especially for lower-performing students – and compresses the grade distribution, eroding the signal value of grades for employers. Gains in AI-specific human capital are offset by losses in traditional human capital, highlighting AI’s possible trade-offs for productivity.

How should higher education systems respond now that every essay-writing student has a free, always-on co-author? Universities are moving in various directions: some prohibit ChatGPT outright while others try to integrate it into curricula. Evidence suggests that AI is being quickly adopted in the workforce to boost productivity, but that its effects in education are mixed. Large-scale field experiments in firms show sizeable productivity boosts, especially for less-experienced workers (Brynjolfsson et al. 2023, Noy and Zhang 2023). A recent VoxEU column estimates that between 1% and 8% of US work hours are already assisted by generative AI, with plausible annual labour-productivity gains of up to 2% (Bick et al. 2024). A recent experimental study finds that participants randomly assigned to practice writing with an AI tool that exposed learners to high-quality, personalised examples performed better on subsequent writing tasks than those not given access to the tool (Lira et al. 2025). Yet there is a real concern that easy access to generative AI crowds out the deep thinking universities are meant to cultivate (Bastani et al. 2024).

What We Ddo

Our new study (Hausman et al. 2025) tracks 36,000 students across 6,000 courses at a leading Israeli university from 2018 to 2024. We exploit the fact that some courses base most of the final mark on take‑home essays or projects (AI‑compatible), while others rely almost exclusively on in-person, supervised exams or lab work (AI‑incompatible). Because students take both types of courses, we can compare a student’s own performance in more versus less AI-compatible courses, after versus before OpenAI released ChatGPT for widespread use in November 2022.

Student and semester fixed effects absorb individual unobserved ability and time trends, and propensity-score matching analysis ensures a tightly controlled comparison of similar-on-observables courses that nevertheless differed in their compatibility with AI. Data in years leading up to the 2022 AI release confirm that grade trajectories in the two course types were parallel pre‑AI. Our design therefore isolates the causal impact of generative AI availability on student performance. We specifically focus on the impact of this tool on both basic and AI-specific human capital development and on the signal value of grades to future employers.

Key findings

  • Grades rise – especially at the bottom. In 2022‑23, average grades in AI‑compatible courses rose by one point on a 0‑100 scale, growing to 1½ points in 2023‑24. The 25th‑percentile student gained two to three points. Failure rates fell by one‑third.
  • The distribution is compressed. Both tails shrink: there were fewer fails, but also fewer low‑pass marks, while the share of very high grades ticks up. Compression weakens the signal value of grades.
  • AI practice yields AI skill but may crowd out fundamentals. Students first exposed to AI in Year 1 perform better in later AI‑compatible courses than similar students for whom AI was not available in Year 1, suggesting genuine acquisition of AI‑specific human capital. Yet they do no better – and sometimes worse – in advanced exam‑based courses, implying substitution away from traditional learning.
  • Within‑student rank predictability falls. Pre‑ChatGPT, a student’s rank in exam courses predicted her rank in take‑home courses; post‑ChatGPT, that slope flattens. Performance with AI reveals less about performance without it.

Why It Matters

Taken together, our results in the classroom echo the broader pattern of rapid, uneven benefits documented in the workplace (Bick et al. 2024). Generative AI lifts struggling students and narrows achievement gaps, but some of the apparent progress comes from outsourcing cognitive effort to the machine. Universities, employers, and policymakers therefore face a twin challenge: teaching students to use AI productively while ensuring they still master core concepts and thought processes unaided.

Policy Recommendations

  • Rethink assessment design. Retain a mix of in-person, supervised exams (to measure individual mastery) and take‑home tasks redesigned to reward crafting effective prompts, critically evaluating AI output, and disclosing tool use.
  • Make AI literacy part of the curriculum. Blanket bans are ineffective and risk widening inequities. Instead, teach students to verify accuracy, detect hallucinations, and attribute sources – skills already demanded in the labour market.
  • Broaden the signaling bundle. As grades lose resolution, employers may rely more on structured interviews, work‑sample tests, or portfolios that demonstrate both foundational knowledge and AI fluency.
  • Monitor equity effects. Because weaker students gain the most, well‑calibrated AI policies could narrow educational inequality. But if reliance on AI erodes fundamentals and is not a complete solution for all tasks, the long‑run productivity gains predicted by macro models may be overstated and effects on inequality are unclear (Acemoglu 2024).
  • Increase focus on intrinsic motivation. Students may need to experience under-preparedness due to overreliance on AI. Finding ways to increase their motivation and sense of responsibility for their own learning may be the key to success.

Conclusion

Generative AI is already reshaping what – and how – students learn. Our evidence shows that easy access to ChatGPT boosts course grades but blurs the information grades convey about underlying skills. The most productive graduates of the AI era will combine robust domain knowledge and critical thinking with the ability to deploy the machine judiciously.

See original post for references

Print Friendly, PDF & Email

16 comments

  1. leaf

    “Once men turned their thinking over to machines in the hope that this would set them free. But that only permitted other men with machines to enslave them.”
    Looks like Frank Herbert was really on to something but what are the odds we get our own Butlerian Jihad?

    Reply
  2. Terry Flynn

    I kinda hesitate to comment on this because I think I’m getting a reputation as a stuck record. However, it resonates with me, but even more seriously, use of “learning for the test” and other short-cuts long pre-date AI and created a generation of clinicians who, quite frankly, I don’t trust.

    You need a basic set of tools in your head, that are as intrinsic as muscle memory, in case we get an EMP or something that makes everyone stop relying on tech for months at a time.

    Again, I am reminded of the store assistant (late teens in *that* major Westfield mall, Sydney) who admired my dual timezone watch but admitted the analogue one was no use to her because she couldn’t tell time. WTAF? *sigh* /oldfartmode

    Reply
  3. Acacia

    This has been discussed previously here at NC, and I agree with a number of others in the commentariat. I also teach university students. My policy on AI tools is simple: no use whatsoever is acceptable.

    Even if it were acceptable, the students must cite it as a source. Failure to do so is plagiarism, which is already deemed unacceptable in higher education. Next, my policy is that AI tools are never an acceptable secondary source, just as Wikipedia is not an acceptable secondary source.

    Just as there have always been problems with plagiarism, there will be students who try to cheat by using AI. If I detect this, they get an automatic “F” for the assignment and possibly for the course — again, it’s no different than plagiarism. The general university policy is that incidents of plagiarism should be reported, and repeated infractions may lead to dismissal from the academic program.

    As this article suggests, there are people who try to integrate “AI skills” into their pedagogy. To this, my feeling is: what skills? Fiddling around with your prompt for an hour to get the desired results — is that a skill? Learning the syntax of an LLM prompt is far simpler than SQL, HTML, or even JSON. This is only a ‘skill’ in the most dumbed-down curriculum imaginable. Moreover, any ‘skills’ that you learn to massage a prompt will evaporate as soon as the model is changed, which is happening constantly. The people who are calling this a ‘skill’ seem to fundamentally misunderstand the technology.

    My assumption is that undergraduate students are in the program to learn how to write and think critically. It won’t happen if they outsource that task to some app. If they don’t care about this, cheat their way through undergraduate study, and only see the degree as a ticket to some corporate job in the future, they will have nothing to offer and might as well be replaced by AI anyway. Graduate students should be learning how to do serious, original research. I see no reason for them to be using AI tools.

    If students want to gamble with their future, that’s their choice.

    Reply
    1. Afro

      I taught a large class this spring. I felt the University did not give me enough help to fight AI and cheating. I only had two proctors for the exams, so it would have been easy for students to cheat. I also think the university should consider setting faraday cages around so e exam rooms.

      Reply
      1. Acacia

        Yes, I hear similar from colleagues and empathize with your situation. I have a friend who quit teaching at one school over this very issue. Many universities are not implementing any policy and basically leaving everything up to the faculty. “Guidelines” are offered, but it’s often more about getting us to take responsibility for dealing with the problem. The only way we can push back is by reducing class size, but often that is out of our control. I’m afraid it’s going to take some larger crisis to compel admins to take this matter seriously.

        Reply
    2. NotTimothyGeithner

      The rising college sophomores experienced shutdown in 8th grade. It’s reasonable to surmise a high portion have simply never done the work expected of a college bound high school student, then matriculated, and wound up in environments where they don’t belong without the skills to even be there. The local colleges are holding classes for freshmen with life skills such as remember to flush.

      Reply
  4. Aurelien

    Sorry, but I’m an absolutist on this question. AI has no place in learning, and should not be allowed. That’s it. There is no “balance to be struck,” and there is no “judicious” use of AI. In the end, AI is getting someone else to do the work for you, and claiming the result as yours. Yes, it’s tedious finding sources, weighing and judging them, assimilating them, coming to conclusions and so forth, but it’s actually a very large part of what you learn during a university course or it should be. Having an LLM make a summary for you of an academic issue with main arguments and references means that you have sub-contracted a very important part of the work to a machine, even if all of the words you use in the final essay are yours. It’s the equivalent of having a Ph D student India do the research for you and writing up the results as your own. In time, and perhaps very quickly, skills of research and analysis will simply disappear, and students will be hostage to a computer programme whose results they just have to accept because they won’t have the skills to check them. And when students brought up on AI go on to become lecturers in ten year’s time, what happens then? Does AI write their lecture notes for them?

    Reply
    1. david

      There was an article in the Guardian in the last year talking about academics use of AI. Basically you had some using it to write course notes, some using it to write assessments and exams and others using it to mark assessments and tests. And of course some academics using it for all the above.

      To me this seems to have resulted in the following:

      Course notes created by AI – course notes read by student’s AI – exam/assessment set by AI -exam/assessment answered by AI – exam/assessment marked by AI

      And it made me think in a few years the human being will be completely irrelevant to academia. It will basically be AI talking to AI with some humans paying fees and others earning wages. But absolutely no transfer of knowledge from human being to human being.

      Reply
      1. Jesper

        It seems that way to me too. AI to human to AI…. The efficient way of dealing with that would be to cut out the middleman (the human) and let AI talk to AI. But I suppose the middleman provides valuable services.

        For students then it seems that universities are more and more seen as places to register at to get a degree and possibly to get valuable connections.
        For people working at universities then it is about providing degrees.
        For government officials then universities is about hiding unemployment and also a way of blaming individuals for government failure in dealing with changing circumstances (job-market).

        Learning, teaching and research are no longer priorities at universities

        Reply
        1. david

          Not even any point having the AI talking to the AI as it adds nothing to the dataset.

          What you describe about universitiesis pretty much how I felt when I studied engineering in the late nineties. Must be even worse now.

          Reply
  5. Adam1

    I’m mostly against using AI for anything where the output is relying on the AI to provide conclusions. I’d rather fact check myself than someone else and with AI you need to do that because of hallucinations.

    That said, I do see places for AI when kept in a nice little box where errors are likely to be rare and likely to be more noticeable; like writing chunks of code. I work with massive amounts of data analytically and therefore have learned to code SQL and R to support my work. If I was allowed to use AI at work I’d definitely consider letting it write some of my code as a lot of time can be consumed learning new packages for a one-off data need or digging through old work to find the right way to structure code I use once every 3 years.

    Reply
  6. matt

    okay so i am an engineering student. this translates to almost none of my classes being ai-compatible. even when i do have a class that involves a project, ai writes terrible papers. trust me man. i tried using ai on a paper last semester and the only way i could get useful information from it was if i fed it papers and told it to only use the papers i fed it as reference material. even then, it failed to connect across topics. AI also cannot do my homework and a better way of cheating on homework will always and forever be getting the answers from your friends who took the class last year.
    a lot of my professors are extremely pro AI. part of this definitely stems from how they use AI in their research – for predicting configurations of molecules or interpreting spectra or whatever. one time a professor told me to use chatgtp to help with a code based assignment. chatgtp was wrong but very helpful in introducing me to functions i wouldn’t have utilized myself.
    im not rlly one way or the other on AI. like it does have its uses. it can be a helpful editing tool. when writing code it’s like if stack overflow and office hours had a baby. when writing emails where i need to sound fake it is also useful. and i know the main reason why i get to be more pro AI is because i am in upper level AI-incompatible courses where bespoke AI for specific tasks exists. and im sure things are different in a 100 level english course.

    Reply
  7. The Rev Kev

    Going forward, you may end up with two types of Universities – those that allow and maybe even encourage the use of AI and those that shut it down entirely. Big businesses are already discriminating against graduates from elite schools as being sub par in performance and attitude so I would also expect big business to only hire from those universities that disallowed AI as they would be wondering just what sort of education a student in an AI-friendly university actually got.

    Reply
  8. Rip Van Winkle

    A well-known Chemistry prof / all subjects podcast guest from upstate NY said a couple of years ago that with AI there is little point to having English composition classes anymore. It’s all going to come back as AI speak.

    I remember when using Cliffs Notes was thought to be a cheat / short-changing yourself.

    Reply

Leave a Reply

Your email address will not be published. Required fields are marked *