Beyond the hype: The practical and ethical implications of generative AI in education
Yan
Imagine a world where a machine could offer students personalised feedback, generate content tailored to their needs, or even predict their learning outcomes.
With the rapid emergence of generative AI, notably the likes of ChatGPT and other large language models (LLMs), such a world seems to be on our doorstep.
However, as the horizon of education broadens with these advancements, we must also consider the maze of ethical challenges that lie ahead.
Educational research has seen accelerated growth in its relationship with LLMs, as evidenced by our new scoping review.
These studies revealed that LLMs have found their way into a staggering 53 types of application scenarios in the automation of educational tasks. These range from predicting learning outcomes and generating personalised feedback to creating assessment content and recommending learning resources.
While this paints a vivid picture of the vast potential LLMs offer in reshaping educational methods, the challenges are many.
Many of the current innovations utilising LLMs have yet to be rigorously tested in real-world educational settings.
Plus, the transparency surrounding these models often remains confined to a niche group of AI researchers and practitioners.
Read more: Tailoring university assessment in the age of ChatGPT
This insularity raises valid concerns about the broader accessibility and utility of these tools in the educational sphere.
Issues of privacy, data usage, and the looming costs associated with commercial LLMs such as GPT-4 add layers of complexity to this discussion.
Beyond the financial concerns, the ethical ramifications of how student data is handled, the potential for algorithmic biases in educational recommendations, and the erosion of personal agency in learning decisions also present significant challenges to widespread adoption.
One can’t help but ponder whether these technologies are primed for widespread educational adoption, or are they reserved for those who can navigate the intricacies of AI and afford the associated costs?
Implications in educational technologies
From our review, three central implications emerge:
Firstly, while there exists a golden opportunity to harness state-of-the-art LLMs for pioneering advancements in educational technologies, it’s imperative to use them judiciously.
Innovations in areas such as teaching support, assessment, feedback provision, and content generation could transform the educational landscape, potentially reducing the burden on educators and enabling more personalised student experiences.
But the economic implications of commercially-driven models like GPT-4 might make this vision more of a dream than a reality.
Secondly, there's a pressing need to elevate reporting standards within the community. In an era dominated by proprietary AI technologies such as ChatGPT, transparency isn’t just a lofty ideal, it's a necessity.
To foster trust and facilitate wider adoption, it’s paramount that we advocate for open-source models (for example, Llama 2), detailed datasets, and rigorous methodologies.
This isn’t merely about boosting replicability – it's about engendering trust and ensuring the tools we advocate for align with the educational community’s broader needs.
Lastly, but by no means least, is the urgent call to adopt a human-centric approach in developing and deploying these technologies. Ethical AI isn’t merely about sticking to a checklist of principles – it’s about weaving human values into the very fabric of these systems.
Stakeholder engagement is key
Engaging stakeholders, from teachers and policymakers to students and parents, in the process of developing, testing, and refining AI technologies ensures that the technology serves the community, rather than the other way around.
When these systems make decisions that impact real lives, those affected should not only be aware, but should have a deep understanding of the rationale, potential biases, and associated risks.
Read more: Rising from the ashes: Higher education in the age of AI
In the end, we think generative AI and LLMs, with their tantalising capabilities, are a double-edged sword. They promise to revolutionise education, but come with a fresh set of challenges concerning ethics, transparency, and inclusivity.
As these models steadily weave themselves into our educational fabric, an active, continuous dialogue among all stakeholders is crucial.
In navigating this brave new world, we must ensure that technological advancements are both ethically sound and genuinely beneficial, leading us not just into the future of education, but a brighter future for all.
This article first appeared as a BERA blog.
About the Authors
-
Lixiang yan
Research Fellow, Centre for Learning Analytics, Faculty of IT
Lixiang Yan is a research fellow at the Centre for Learning Analytics at Monash University. His current research focuses on the application of multimodal learning analytics and generative artificial intelligence in supporting educational research and practice. He also has extensive experience in developing and integrating educational technologies into authentic learning environments.
Other stories you might like
-
Episode 86: Are Humans About to Evolve?
Trace the increasingly blurred line between man and machine in the world of transhumanism on our “What Happens Next?” podcast.
-
AI, we need to talk: The divide between humanities and objective truth
Could our fascination with objectivity be the Pied Piper that led us to develop a machine some of us now fear and avoid?
-
Tailoring university assessment in the age of ChatGPT
The immediate challenge for educators is to determine what an AI-literate skill set looks like, in order to continue to teach and assess the core skills that have traditionally sat at the heart of the university.
-
So sue me: Who should be held liable when AI makes mistakes?
AI is more likely to make errors than humans, because it relies on often incomplete or inaccurate data. The big question is, who’s accountable if it does? The user, programmer, owner, or AI itself?