Beyond the hype: The practical and ethical implications of generative AI in education

Imagine a world where a machine could offer students personalised feedback, generate content tailored to their needs, or even predict their learning outcomes.

With the rapid emergence of generative AI, notably the likes of ChatGPT and other large language models (LLMs), such a world seems to be on our doorstep.

However, as the horizon of education broadens with these advancements, we must also consider the maze of ethical challenges that lie ahead.

Educational research has seen accelerated growth in its relationship with LLMs, as evidenced by our new scoping review.

These studies revealed that LLMs have found their way into a staggering 53 types of application scenarios in the automation of educational tasks. These range from predicting learning outcomes and generating personalised feedback to creating assessment content and recommending learning resources.

While this paints a vivid picture of the vast potential LLMs offer in reshaping educational methods, the challenges are many.

Many of the current innovations utilising LLMs have yet to be rigorously tested in real-world educational settings.

Plus, the transparency surrounding these models often remains confined to a niche group of AI researchers and practitioners.

This insularity raises valid concerns about the broader accessibility and utility of these tools in the educational sphere.

Issues of privacy, data usage, and the looming costs associated with commercial LLMs such as GPT-4 add layers of complexity to this discussion.

Beyond the financial concerns, the ethical ramifications of how student data is handled, the potential for algorithmic biases in educational recommendations, and the erosion of personal agency in learning decisions also present significant challenges to widespread adoption.

One can’t help but ponder whether these technologies are primed for widespread educational adoption, or are they reserved for those who can navigate the intricacies of AI and afford the associated costs?

Vector illustration with the 'LLM' and 'large language model' above a brain with coloured wires dangling beneath

Implications in educational technologies

From our review, three central implications emerge:

Firstly, while there exists a golden opportunity to harness state-of-the-art LLMs for pioneering advancements in educational technologies, it’s imperative to use them judiciously.

Innovations in areas such as teaching support, assessment, feedback provision, and content generation could transform the educational landscape, potentially reducing the burden on educators and enabling more personalised student experiences.

But the economic implications of commercially-driven models like GPT-4 might make this vision more of a dream than a reality.

Secondly, there’s a pressing need to elevate reporting standards within the community. In an era dominated by proprietary AI technologies such as ChatGPT, transparency isn’t just a lofty ideal, it’s a necessity.

To foster trust and facilitate wider adoption, it’s paramount that we advocate for open-source models (for example, Llama 2), detailed datasets, and rigorous methodologies.

This isn’t merely about boosting replicability – it’s about engendering trust and ensuring the tools we advocate for align with the educational community’s broader needs.

Lastly, but by no means least, is the urgent call to adopt a human-centric approach in developing and deploying these technologies. Ethical AI isn’t merely about sticking to a checklist of principles – it’s about weaving human values into the very fabric of these systems.

Stakeholder engagement is key

Engaging stakeholders, from teachers and policymakers to students and parents, in the process of developing, testing, and refining AI technologies ensures that the technology serves the community, rather than the other way around.

When these systems make decisions that impact real lives, those affected should not only be aware, but should have a deep understanding of the rationale, potential biases, and associated risks.

In the end, we think generative AI and LLMs, with their tantalising capabilities, are a double-edged sword. They promise to revolutionise education, but come with a fresh set of challenges concerning ethics, transparency, and inclusivity.

As these models steadily weave themselves into our educational fabric, an active, continuous dialogue among all stakeholders is crucial.

In navigating this brave new world, we must ensure that technological advancements are both ethically sound and genuinely beneficial, leading us not just into the future of education, but a brighter future for all.

This article first appeared as a BERA blog.