Don’t Encourage People to Run Before They Can Walk: Rethinking Some GenAI Paradoxes

By Junhong Xiao, Emeritus Professor, Open University of Shantou and David Lim, Director, Centre for Digital Education Futures (CENDEF), Open University Malaysia

Recent debates on the use of generative AI in research, teaching, and learning have increasingly framed AI assistance through the lenses of transparency and intellectual humility. Within this framing, the use of AI tools is often presented as a legitimate response to uncertainty or gaps in one’s knowledge, provided that such use is disclosed and undertaken in good faith. Admitting “I’m not sure” and seeking assistance is rightly seen as a professional virtue rather than a failing.

While this position is intuitively appealing, it risks obscuring a more fundamental issue: the situated appropriateness of AI use. The question is not merely whether AI assistance is transparently acknowledged or motivated by humility, but whether its use is appropriate given the user’s level of intellectual and professional readiness. AI tools are not universally benign aids. In some contexts, their use may undermine rather than support sound judgement.

At the heart of this issue lies a simple but often overlooked paradox. To verify, evaluate, or improve AI-generated outputs in a given domain, one must already possess sufficient expertise in that domain. This is not a controversial claim but a matter of common sense. Current AI systems, including the most advanced generative models, are well known to produce errors, omissions, and fabrications. Their outputs therefore require meaningful human oversight. That oversight cannot be performed adequately by someone who lacks the relevant knowledge and experience.

This paradox becomes especially clear when considering the use of AI tools for academic writing and research. It is frequently suggested that researchers who are less confident in a topic or in academic writing can rely on AI to identify literature, generate summaries, or improve language, provided they “check the sources” or “take responsibility for the final output.” Yet this raises an obvious question: if a researcher does not already have sufficient expertise in the topic or the language, how can they reliably judge the accuracy, completeness, or quality of what the AI produces? If they are able to evaluate the output confidently, one might reasonably ask what substantive problem the AI is solving in the first place.

The same paradox becomes even more visible in teaching and learning contexts. Advocates of generative AI in teaching sometimes argue that traditional skills such as writing essays or literature reviews are becoming less important, and that students should instead be taught to analyse, critique, and improve AI-generated texts. However, this proposal collapses under scrutiny. If students do not already know how to write a coherent essay or literature review themselves, they are not in a position to judge whether an AI-generated version is any good, let alone to revise and improve it meaningfully. One cannot critically evaluate a practice one has never mastered.

"AI tools are not universally benign aids."

A similar logic applies to teachers’ use of AI-generated lesson plans. Lesson planning is a core professional responsibility, central to pedagogical judgement and instructional quality. If a teacher is capable of designing sound lesson plans, the value of outsourcing this work to AI and then spending time correcting and refining the output is questionable. If a teacher is not capable of designing sound lesson plans, then they are not in a position to evaluate or improve AI-generated ones. In this case, the issue is not efficiency but professional competence. The uncritical use of such tools risks contributing to the erosion of pedagogical autonomy and professional judgement.

None of this is to deny the powerful affordances of generative AI for research and education. Nor is it to argue against personal experimentation or individual choice. People may find AI tools useful or unhelpful in different ways, and those decisions can reasonably remain personal. The problem arises when individual experiences are presented, implicitly or explicitly, as general recommendations. Framing AI use as universally appropriate risks doing a disservice to those who are not yet intellectually or professionally prepared to engage with these tools critically and responsibly.

"generative AI does not eliminate the need for foundational skills, disciplinary knowledge, or professional competence."

Transparency and intellectual humility are important scholarly virtues. However, they are not sufficient conditions for responsible AI use. What matters at least as much is expertise, understood not as elitism but as the capacity to exercise informed judgement. Without such capacity, claims of responsible or ethical AI use become fragile.

The broader lesson is straightforward: generative AI does not eliminate the need for foundational skills, disciplinary knowledge, or professional competence. On the contrary, it amplifies their importance. Encouraging people to rely on AI before they have learned to walk on their own risks confusing assistance with understanding, and convenience with competence. In discussions of AI adoption in research and education, greater attention must therefore be paid not only to how AI is used, but to who is ready to use it, in what contexts, and with what forms of responsibility.