Published May 15 2023

Tailoring university assessment in the age of ChatGPT

OpenAI has just released ChatGPT-4, the latest iteration of its powerful AI text generator. The tool is capable of generating convincingly human-like responses to almost any question users put to it.

It can write limericks, tell jokes, and plot a novel. And it can draft a convincing response to almost any question a high school teacher or university lecturer might ask students to write about.

Previous iterations of OpenAI’s GPT language models would often generate text riddled with strange and obvious mistakes. ChatGPT (and particularly GPT-4) is less prone to these problems. The responses that ChatGPT generates are apparently capable of passing exams across many disciplines.

It’s tempting to think we’ll always be able to distinguish between the work of AI and the work of humans, particularly when it comes to distinctly “human” tasks such as creative writing, careful reasoning, and drawing novel connections between different kinds of information.

Unfortunately, this optimism is misguided. AI-generated prose and poetry can be beautiful. And with some clever prompting, AI tools can generate passable argumentative essays in philosophy and bioethics.

This raises a serious worry for universities that students will be able to pass assessments without writing a single word themselves – or necessarily understanding the material they’re supposed to be tested on.

This isn’t just a worry about the future; students have already begun submitting AI-generated work.

Cheating bans hard to enforce

Some institutions treat the use of AI text-generators as “cheating”. For example, public school systems in NSW and Victoria have banned ChatGPT.

But such bans will be hard to enforce. Compared to traditional forms of plagiarism (where plagiarised work can be compared directly with the source material), student use of AI-generated text is hard to detect – and harder still to prove, in part because new ChatGPT generates new responses each time a user inputs the same prompt.

One might hope the problem is just temporary. For its part, OpenAI (alongside other developers) is developing tools to detect AI-assisted “cheating” – though such tools are prone to making mistakes, and can at present be circumvented by asking ChatGPT to write in a style that its own detector is unlikely to catch.

OpenAI is also developing a technique for “watermarking” AI-generated text. This works by influencing which words the AI chooses, leaving subtle quirks in the writing that can be detected further down the track.

But this strategy also has serious limitations. Lightly editing AI-generated text would presumably overwrite these watermarks. And students could seek out alternative AI tools that don’t watermark AI text.

The educational challenges of AI text generation are here to stay.

Adapting won’t be an easy process

Generative AI tools such as ChatGPT are poised to make far-reaching changes to how we approach writing tasks.

These tools are exciting. Among other things, they’ll make some tedious and difficult parts of the writing process easier (including for students for whom English is a second language).

But adapting to a world where AI can write compelling student essays will not be straightforward.

Sam Altman, the CEO of OpenAI, has compared the release of ChatGPT to the advent of the calculator. Calculators brought about enormous benefits; ChatGPT will, Altman claims, do the same.

Schools have adapted to calculators by changing how maths is tested and taught; we now need to do the same for ChatGPT.

This analogy seems more or less right, but its implications are less reassuring than Altman intends. The task of adapting to calculators has been long and arduous, and it remains unclear how best to incorporate them into the classroom.

Mathematicians continue to raise plausible worries that overuse of calculators impedes the development of important forms of “mathematical thinking”.

Rather than comforting us, the parallel with calculators should alert us to the magnitude of the task we face.

The two main threats

We see two main threats posed by tools like ChatGPT.

The first is that they’ll produce content that’s superficially plausible but entirely incorrect.

AI outputs can thus leave us with a deeply mistaken picture of the world.

Contrary to appearances, ChatGPT is not “trying” (but, often, failing) to assert facts about the world. Instead, it is (successfully) performing a different task – that of generating superficially plausible or convincing responses to a prompt.

In humans, we would describe this kind of behaviour as “bullshitting”. The term also aptly describes many AI outputs.


Read more: ChatGPT: We need more debate about generative AI’s healthcare implications


The second worry is that reliance on these tools will result in the erosion of important skills. Essay writing, for example, is valuable in part because the act of writing can help us think through difficult concepts and generate new ideas.

Good arguments often emerge in an iterative fashion over several drafts, where the final version looks quite different to what one envisaged in their original plan.

Writing, so understood, is not just an act of translating our thoughts onto the page; for many writers, it’s part of the reasoning process. And if we begin to outsource the writing process to artificial intelligence, it’s a skill students might lose if they come to rely on AI text generators.

Addressing the challenges

How, then, should we meet these challenges? In these early stages of the introduction of generative AI, educators may feel overwhelmed by the rapidly changing technological environment, but students are also coming along for the ride with us.

Being transparent about decisions we make regarding when and how AI tools are used in assessment could help ease anxiety on both sides.

We suggest four approaches.

Teach students how (not) to use AI text generators

ChatGPT can be a useful tool. It can, for instance, help generate ideas and get words on the page.

The worries about misinformation are serious. But these are best addressed by teaching students how to use these tools, how to understand their limitations, and how to fact-check their output.

Fortunately, the core skills cultivated by a good education provide a strong foundation for this project. Teaching students how to read critically, how to evaluate or corroborate evidence, and how to distinguish good arguments from bad, are things universities should be doing already.

One approach might be to develop specific assessment tasks where students generate, analyse, and criticise AI outputs.

While such tasks might have some role to play, we would caution against placing generative AI at the centre of education. There’s no shortage of bad arguments being put forward by human beings. We should teach students to recognise bullshit wherever it appears.

Continue efforts to design ‘authentic’ assessments

We should remind ourselves that for most, choosing to participate in higher education comes from a genuine interest in a subject.

This fact may go some way towards mitigating the temptation to outsource their studies to AI, particularly when the value of completing this work is clear to students.

By designing assessments that are relevant to students’ future careers, and clarifying the purpose of tasks in relation to their own development, we can encourage learners to engage with assessment in the way we intended.

Assessment that engages with, and leverages, students' interests could motivate learners to remain engaged such that they don’t see value in outsourcing the pursuit of their own knowledge to AI.

This is a desirable way of designing assessment, both as a response to ChatGPT and as a general goal for higher education.

But while this strategy may undercut some students’ desire to turn to AI, it’s unlikely to fully dissolve the temptation to save time by having an AI write your essay – particularly where doing so would not lower (and might even raise) your final grade.

Additional strategies are needed.

Balance essays with other types of assessment

A key worry about AI text-generation is that students won’t actually understand what they appear to, given the work they’ve submitted.

This concern can be met by balancing written work with other kinds of assessments.

In particular, in-person oral presentations cannot be taken over by any algorithm, and so may be an ideal option (provided, of course, that any increase in workload for teaching staff is supported by the institution).

Supplementing traditional essays with other assessments need not come at the expense of good assessment design.

To the contrary, there are good educational reasons to vary written work with these other kinds of assessment; oral communication skills are enormously valuable across a range of professions.

At the same time, writing skills are poised to become less important as ChatGPT and tools like it start to play a major role in how text is produced, both within and outside the classroom.

Develop AI-resistant written assessments

Another strategy involves designing assignments where students are either required to demonstrate their own understanding (independently of written work they’ve submitted), or where AI-generated bullshit can be easily detected.

The revival of pen-and-paper exams is being considered by Australian universities.

This strategy may have a role to play, but it would come at a cost. We’re amid a shift away from pen-and-paper examinations to “authentic” assessments – that is, assessments that evaluate skills students will employ in real-world settings.

Few workplaces require their employees to write detailed discussions of difficult questions by hand, in isolation, and without the ubiquitous modern conveniences of an internet connection and a word processor.


Read more: Rising from the ashes: Higher education in the age of AI


An alternative is to combine written essays with the presentation and discussion of this work during class time, potentially modelled after the format of a viva presentation or thesis defence (albeit made gentler and shorter according to the cohort being taught.)

The viva format requires students to understand, and be able to communicate, the ideas they’ve defended in their essay – regardless of whether they outsourced any of the writing to an algorithm.

Finally, there’s some scope to develop essay questions that ChatGPT is unlikely to perform well on.

In our own experiments, we found ChatGPT can generate convincing responses about major works in our respective disciplines.

However, it fares very poorly when asked about the cutting edge of scholarly debate, since the corpus of work it was trained on contains much less (if any) discussion of this work.

When asked to reference its claims, it’s prone to “hallucinate” sources that don’t exist.

Assignments that ask for deep research on recent scholarly developments are, for now, relatively AI-resistant. But this strategy will need to be continuously reviewed as AI language models continue to improve.

Picking the right tool for the job

Assessment regimes should be tied to the educational goals we’re trying to achieve, not a knee-jerk reaction to new technologies.

We recommend that educators considering how to adapt their assessment regime in light of ChatGPT ask themselves the following questions:

 

These suggestions differ from some existing responses, which centre on detecting or preventing the use of AI tools.

This is for a good reason. Our students will enter a world where the use of such tools will likely be commonplace. Indeed, ChatGPT has already been listed as an author on recent academic papers, and utilised by journalists and authors.

We need to prepare students for a world where tools like ChatGPT are readily available – and to teach them how to use these tools well.


Read more: ChatGPT: Old AI problems in a new guise, new problems in disguise


In the longer term, progress in generative AI seems likely to pose even more profound challenges to educators.

In the not-too-distant future, ChatGPT and tools like it may become better than most humans at writing in many genres and on many topics.

It’s far from clear whether students will bother to learn to write in such circumstances – or whether employers will care whether they do.

Given the connection between learning to write and learning to think, this is a disheartening prospect.

Moreover, it’s almost certain generative AI will become capable of tasks that many educators think of as central to their own profession. ChatGPT can already generate useful feedback on written work.

Non-profit education organisation Khan Academy recently announced it will be launching a GPT-4 powered AI-assistant, Khanmigo, which can play the role of a tutor to students, and can also assist teachers with tasks such as creating lesson plans.

Dystopian visions in which AI teachers set tasks that students then farm out to AI look all too plausible.

The immediate challenge for educators is to determine what an AI-literate skill set looks like, and how to evaluate whether students have these skills, especially when many of us are new to these skills ourselves.

If we can get this right, perhaps we’ll be able to continue to teach and assess the core skills of research, reflection, argument, and critical thinking that have traditionally sat at the heart of the university.

The deeper challenge posed by the “threat” of AI is to imagine what education would look like should the tools available to us relieve us of the need to exercise these crucial skills.

 

About the Authors

  • Julian koplin

    Lecturer, Monash Bioethics Centre

    Julian is a lecturer in bioethics. His research interests include human tissue ethics, stem cell ethics, and the ethical implications of new technologies.

  • Robert sparrow

    Professor of Philosophy, Monash Data Futures Institute

    Robert's research interests are bioethics, political philosophy and applied ethics; he is an expert in philosophical arguments with real-world implications.

  • Joshua hatherley

    PhD Candidate, School of Philosophical, Historical, and International Studies

    Joshua is a PhD candidate in the School of Philosophical, Historical, and International Studies. His research projects are concerned with identifying and responding to ethical trade-offs in the design, use, and implementation of emerging healthcare technologies, particularly machine learning and AI.

  • Nicola rivers

    Senior Teaching Fellow, Obstetrics and Gynaecology Monash Health

    Nicola is a reproductive biologist and lecturer at Monash University. Having completed a Bachelor of Biomedical Science and postgraduate study in reproductive biology, her main interest lies in the application of assisted reproductive techniques in conservation and species management. Specifically, Nicola’s research focuses on the development of reproductive technologies for the preservation of endangered fish species in Australia. She’s currently a senior teaching fellow within the Education Program for Reproduction and Development at Monash. In addition to her teaching role, Nicola is part of the management team for the Australian Frozen Zoo, a biobanking initiative responsible for collecting and storing cells from some of Australia’s unique wildlife.

  • Iwan williams

    Former PhD Candidate, Cognition and Philosophy Lab, Faculty of Arts

    Iwan conducts research in the philosophy of mind and cognitive science. His PhD focuses on the “format” of cognitive representations, and aims to demonstrate that distinctions in representational format prove exceptionally useful tools in carving up the conscious and unconscious mind.

Other stories you might like