Specialist Language Courses

Beyond Vocabulary Lists – What AI Writing Tools Really Mean for Medical English

Human vs AI in Healthcare

Already in the room

The question is no longer whether AI tools will enter the medical English writing classroom. They already have. The more pressing question is how to integrate them purposefully, without sacrificing the development of genuine language skills or undermining academic integrity. A wave of recent research across medical English, English for Academic Purposes (EAP), and English for Specific Purposes (ESP) is beginning to supply some answers.

The evidence base is strengthening

The medical English case has been building for some time. A 2024 study published in BMC Medical Education examining AI-assisted academic writing among non-native speaking medical students found that students who received AI-assisted instruction showed improved proficiency across multiple aspects of writing, including organisation, coherence, grammar, and vocabulary (Li et al., 2024).¹ Encouragingly for hard-pressed educators managing large cohorts, statistical analysis revealed no significant difference between manual scoring and GPT-4 scoring, pointing to the potential of large language models to assist teachers in the grading process (Li et al., 2024).¹

The English for Academic Purposes (EAP) literature now adds considerable weight to these findings. A 2026 study published in Discover Education analysed postgraduate non-native English speakers enrolled in a mandatory EAP writing course. It found that generative AI-assisted writing feedback, when used with pedagogical awareness, can effectively enhance EFL learners’ writing skills and support active participation in the writing process. However, the researchers also stressed that critical thinking skills are essential and that students should avoid over-reliance on AI tools (Zhang et al., 2026).² Meanwhile, a 2025 article in ELT Journal exploring structured AI chatbot use in an online EAP module reported that unstructured AI use had led to increased student anxiety and poor learning outcomes, while structured use helped address those challenges. This is a timely reminder that education design matters as much as the tools themselves (Webb & Şenaydın, 2025).³

The feedback question

One of the most practically useful threads in the recent literature concerns the nature and quality of AI-generated feedback. Research synthesised in a 2025 review published by De Gruyter found that teacher feedback tends to focus more on content, while GenAI feedback focuses more on organisation (Xu & Guo, 2025).⁴ This suggests the two work best in combination rather than in competition.

A 2026 study in Social Sciences and Humanities Open comparing types of AI prompts found that EFL-tailored prompts produced the highest quality ChatGPT feedback in terms of relevance, clarity, usefulness, and tone (Algobaei & Alzain, 2026).⁵ The implication for medical English teachers is direct: prompts specifically oriented to clinical register, case report conventions, or patient communication genres are likely to produce more useful feedback than generic grammar-checking instructions.

A framework for responsible integration

For institutions uncertain about where to draw the line between legitimate scaffolding and problematic dependence, a framework published in TESOL Journal in early 2026 offers a practical starting point. The EAP AI Assessment Scale (EAP-AIAS) consists of five levels ranging from ‘No AI’ to ‘Full AI’. Each delineates appropriate GenAI usage in EAP tasks, and frames GenAI as mediation that can extend learner performance beyond what is reliably achievable independently (Tregubova, 2026).⁶ The scale reflects growing recognition that existing EAP curricula may inadequately prepare students to incorporate GenAI tools appropriately into their academic writing practices (Wu et al., 2026).⁷

A parallel framework, published in Artificial Intelligence in Education in 2026, applies the AI Assessment Scale specifically to EFL writing and translation contexts, demonstrating how different levels of AI assistance can be mapped to different stages of task completion and learner proficiency (Roe, Perkins, & Furze, 2026).⁸ For medical English programmes, either framework could be adapted to map AI use against the different genres students need to master, such as clinical case reports, patient discharge summaries, research abstracts, and reflective portfolios – with different levels of AI assistance deemed appropriate at different stages of proficiency and task complexity.

The emerging consensus

The picture that emerges across all of this research is nuanced but actionable. AI works best as a scaffolding tool — helping students draft, revise, and self-edit — while the teacher’s role shifts towards higher-order feedback on register, precision, and clinical appropriateness.

A 2025 study in Frontiers in Artificial Intelligence reviewing external variables influencing students’ attitudes towards AI writing tools found that motivation, engagement, and positive societal expectations are significant factors in determining whether students use these tools effectively – and that educators play a key role in shaping all three (Mansoor, Sumardjoko, & Sutopo, 2026).⁹

For medical English educators, the opportunity is not just to adopt these tools but to shape how they are used. The domain-specific knowledge that EMP specialists bring –  understanding of clinical genres, register variation, the communicative demands of different healthcare settings – is precisely what is needed to ensure that AI writing support in medical education serves real language development goals, rather than simply producing polished text that bypasses them.

References

1. Li, J., Zong, H., Wu, E., Wu, R., Peng, Z., Zhao, J., Yang, L., Xie, H., & Shen, B. (2024). Exploring the potential of artificial intelligence to enhance the writing of English academic papers by non-native English-speaking medical students — the educational application of ChatGPT. BMC Medical Education, 24, 747. https://doi.org/10.1186/s12909-024-05738-y
2. Zhang, Z., et al. (2026). Generative AI-assisted feedback and EFL writing: A study on proficiency, revision frequency and writing quality. Discover Education. https://doi.org/10.1007/s44217-025-00602-7
3. Webb, R., & Şenaydın, F. (2025). Developing students’ agency and voice by using generative AI in an online EAP module. ELT Journal. https://doi.org/10.1080/17501229.2025.2538781
4. Xu, J., & Guo, X. (2025). Generative AI and second language writing. DSLL — Language Learning Research. De Gruyter. https://doi.org/10.1515/dsll-2025-0007
5. Algobaei, F., & Alzain, E. (2026). Prompt engineering for non-native English learners: A generative AI approach to personalised language feedback. Social Sciences and Humanities Open, 13, 102341. https://doi.org/10.1016/j.ssaho.2025.102341
6. Tregubova, Y. (2026). The EAP-AIAS: Adapting the AI Assessment Scale for English for Academic Purposes. TESOL Journal, 17(2), e70122. https://doi.org/10.1002/tesj.70122
7. Wu, C., Moorhouse, B. L., Wan, Y., & Wu, M. (2026). Exploring PhD students’ utilisation of generative AI in academic writing for publication purposes: Insights for EAP. Journal of English for Academic Purposes, 79, 101612. https://doi.org/10.1016/j.jeap.2025.101612
8. Roe, J., Perkins, M., & Furze, L. (2026). Applying the AI assessment scale in writing and translation for EFL. Artificial Intelligence in Education, 2(2), 59–74. https://doi.org/10.1108/AIIE-02-2025-0034
9. Mansoor, H. S., Sumardjoko, B., & Sutopo, A. (2026). External variables influencing the attitudes of students toward AI acceptance in improving English writing: A systematic review. Frontiers in Artificial Intelligence. https://doi.org/10.3389/frai.2025.1719955

Did you find this content valuable?

Subscribe to the Medical English Newsletter to receive monthly updates, insights, and resources on Medical English.

More To Explore

Celebrate International Nurses Day with 10% off Nursing subscriptions!