Resistance is futile. If, like me, you are pondering the future of liberal arts education in the age of AI, these thoughts might have run through your mind. As if our existence in humanities programmes wasn’t challenging enough, a whole new world opened in November 2022 when Open AI released Chat GPT. Since then, there has been an explosion of large language models (LLMs) ready to do your research, summarise the evidence and write your papers. If you are an instructor trying to teach students how to locate and evaluate sources and to read and write critically, this new technology poses a problem, putting it gently.
Total prohibition of LLMs is not viable, however. After all, it is very likely that any writing done on a word processor will have some AI intervention, whether we are cognisant of it or not. Grammarly has integrated an LLM into its interface; Google’s “help me write” feature is an LLM, and Microsoft Word has been telling me how to write for some time (which is really annoying).
Campus resource: Enhance GenAI collaboration for future-proof research support
But I do insist that we make reflective, intelligent decisions about the impact that something like GPT 4 – the much more powerful next iteration of ChatGPT – will have on the intellectual development of our young people. After all, those of us reading this article – not to mention the programmers who built the LLMs – had the opportunity to learn how to think and write critically without mechanical help: an ability that is needed in every profession and is part of what it means to be a thinking human being. The brain is like a muscle; it requires exercise. I fear that before we even know if the use of these LMMs helps or harms student learning, we will see a flurry of highly questionable decision-making in universities.
Many educators will say we must “follow the science”. But when it comes to putting that into practice, they opt time and again for the latest technological innovation over what students need. Numerous studies over the last 25 years show that we learn better and comprehend more when we read print on a page, rather than pixels on a screen (and this is also true for writing on paper over typing). We are far more likely to remember what we have read when we see it, touch it and are not distracted by hyperlinks and all the other magical things we can do with a screen. Yet in the last 10 years we have seen a rush to put iPads in the hands of children as young as five. And we have allowed students to bring their phones to class even though studies have shown that their phones’ mere presence lowers a student’s comprehension.
中国A片
In intellectual, scientific and creative work in particular, future workers will still need to be able to draw on a body of knowledge, read closely and critically, organise their thoughts, find evidence and communicate ideas. Even in professions that will be assisted by some form of AI, humans with a solid understanding of the data and the ability to evaluate evidence will still be needed. This kind of intellectual work is not the work of an empty head.
College is when young people are given the opportunity to develop these skills through the guidance of skilled faculty, such as those teaching for Purdue University’s Cornerstone integrated liberal arts programme – a two-semester, first-year sequence that aims to develop students’ communication and creative thinking skills, broaden their outlooks and cultivate their minds through exposure to transformative texts. If students ask GPT 4 to do their work for them, whether summarising a novel they were supposed to read or writing the essay they were supposed to write, they are not creating the kind of attention and memory necessary for deep thought.
中国A片
There is something else, too, that students need to learn for future citizenship – something?LLMs cannot replicate. That is the ability to understand, tolerate and even appreciate subtlety, ambiguity and nuance. Few things in this world are simply right or wrong, good or evil, but the digital world of fast answers makes us susceptible to propaganda, hyperbole, dogma and venom. If we offload intellectual work to algorithms, our susceptibility is only likely to increase, with dire consequences for our political culture.
Or, alternatively, if we all slavishly absorb what the machines tell us, we?might create one homogenised culture that blurs diversity and identity. To my mind, that would be a catastrophe.
So let’s make smart decisions. If used responsibly by those who already know how to think critically, generative AI might be one tool in a large toolkit – like calculators for mathematicians, as the oft-repeated analogy goes. I hope so. But from what I have seen of our American obsession with every new shiny object regardless of the harm it might do, I worry.
My plea is simply this: insofar as general education is concerned, let’s continue to teach our students the joy of finding their own voices, their individuality and their creativity. Let’s continue to ask them to read actual books, find and evaluate evidence on their own and write their own work.
中国A片
Because, yes, in the future, they will probably work with AI. But if we don’t teach them to think for themselves, the risk is that they one day find themselves working for AI.
is Germaine Seelye Oesterle professor of history and director of the Cornerstone programme at Purdue University.
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to 罢贬贰’蝉 university and college rankings analysis
Already registered or a current subscriber? Login