Wednesday, December 31, 2025

The AI-Powered Reading Test Workflow



月考命題不想用出版社現成的,何不自己來寫學校的活動、故事,然後出得很像國中英語會考的閱讀測驗?這在以前非常花時間、也很有可能會出現不自然的英文。現在有了AI,幾分鐘就可以生成一篇既符合學生程度、又跟他們學校生活直接相關的閱讀測驗,而且保證非常的素養和會考!

Writing your own reading passages? It’s a nightmare. Trying to design question sets that actually mimic the discriminative difficulty of the CAP? Even harder. But I found a workaround, with a workflow that actually works so very effectively.

We all know the problem with textbook publisher materials. They are not relatable to students’ real life or interests. Or worse, students might see the exact same reading passages and tests at cram school three weeks ago. So, we write our own? Sure. That’s what I have done before AI. It’d be really challenging and time consuming, not to mention awkward phrasing. After all, we’re not native speakers.  

Guess what? I ran a workshop recently to fix this. We built a method to turn local school news into CAP-grade tests without the headache, and I was glad to see how the participants react to this approach. Here is the workflow:



1. Smart Translation

Don't start from scratch; start with reality. Grab a school Facebook post about the English Singing Contest. A news snippet about an international postcard exchange. Something they care about and relatable.

Use Gemini or ChatGPT to do the heavy lifting. But don't just say "translate." Use this specific command:

"Translate the following text into English, ensuring it sounds natural and has no awkward phrasing for native speakers. Make it understandable for 2nd graders."

Why? Because it fix the problem for a massive vocabulary list of the word bank. You can adjust the difficulty level before you start making the exam, not after.

2. The NotebookLM Magic

Here is the secret sauce. I’ve tried to create a master prompt with Gemini but it wasn’t doing what I’ve imagined. You have to train it first with real, previous CAP tests as sources on NBLM.

l  Feed: Upload the last three years of official CAP tests into Google NotebookLM.

l  Analyze: Ask the AI to break down the genre, word counts, and logic.

l  Input: Paste your translated text from Step 1.

l  Execute: Tell it to generate a question set that mimics the exact style of the analyzed data.

3. The Human Filter

AI could go wrong. We found an obvious flaw in a question stem, with the intended inferencing phrase totally unnecessary. So, you still have to read it and swap out the obscure words when necessary. The AI gets you 90% of the way there, and you’ll only have to work on the last 10%. Before you hit print, check for grammatical errors, as AI can sound confident yet still hallucinate.

This method helped me build two full reading sets that stood up to strict peer scrutiny by the workshop participants. Combining human intuition with generative tech takes the pain out of the process—and honestly, it gave me the confidence to show other teachers how to do it for next semester’s EAT school-based services. 


No comments:

Post a Comment