
What changed in 2024–2025
For years, many classroom conversations about AI writing tools focused on “autocomplete”: a sentence suggestion here, a paragraph continuation there. Students could accept or ignore prompts, but the tool largely worked at the level of local text. The 2024–2025 shift is different. Many tools now behave more like co-authoring spaces: they can see and act on the whole document, keep track of goals, and revise at scale. In practice, that means students can draft in a canvas-style environment, ask for a rewrite in a different tone, request a tighter argument, or generate alternative structures without starting from scratch.
This matters because the evidence of authorship is no longer found in small traces like odd word choices. Document-aware assistants can smooth those away, producing writing that looks plausible, polished and consistent. The centre of gravity has moved from “Can the tool generate text?” to “Can the student make good decisions with text?” If you want a concrete sense of how drafting spaces are evolving, the patterns described in openai-canvas-drafting-guide mirror what students increasingly experience across platforms, even when the tool has a different name.
Why this matters for writing
When tools become co-authors, the failure modes change. The first is over-polish: writing that is technically correct but strangely flat, with safe phrasing and generic examples. Students may believe they have “improved” their work, yet the piece loses specificity, risk-taking and genuine insight. A second is voice drift. A student starts with a lively, personal opening, then asks the assistant to “make it more formal” and ends up with a different writer on the page. The third is hidden outsourcing: not simply using AI, but letting it make the key choices—argument, evidence selection and structure—while the student becomes a manager of output rather than an author.
These issues are not solved by detection tools. Detection is unreliable, easy to circumvent, and often punishes students whose writing already looks “too polished”, including multilingual learners. More importantly, detection frames the problem as policing. Writing instruction works best when it frames the problem as craft: planning, drafting, revising and sourcing with intention. That is why an evidence-first approach is a better fit for 2024–2025. It shifts assessment from “What did you produce?” to “How did you get there, and why?”
An evidence-first model
An evidence-first model treats AI use as permissible, but requires a trail of decisions. The goal is not to generate paperwork; it is to collect small, meaningful artefacts that show thinking at key moments. In a typical writing cycle, you want evidence of three things: intent (what the student was trying to do), judgement (what they changed and why), and grounding (where claims and examples come from).
In practical terms, that means collecting a prompt log, a revision rationale, and a source trail. A prompt log can be as simple as a short record of the student’s key prompts and the tool’s responses, captured at two or three moments rather than continuously. A revision rationale is a brief explanation of the student’s main changes between drafts, including at least one change they rejected. A source trail shows how information entered the piece: links, page references, interview notes, class texts, or dataset citations. Together, these artefacts make the writing assessable even when the final prose has been heavily edited.
Timing matters. If you only ask for evidence at the end, students will reconstruct a story that sounds plausible. If you collect evidence in small checkpoints, you capture authentic decision-making. This is the same logic behind strong digital citizenship routines: we teach students to document, attribute and reflect as they go, not after the fact. If you are building shared language around responsible use, digital-citizenship-and-ai offers a helpful foundation to align expectations across subjects.
Classroom routines that show thinking
The most workable routines are short, repeatable, and tied to moments when students naturally pause. A five-minute “prompt pause” can happen after planning, after the first paragraph, and during revision. Students simply copy their last two prompts and the assistant’s output into a log, then annotate one line: “What I kept, what I changed, and why.” This turns AI from a hidden helper into a visible drafting partner, and it gives you something assessable without reading every chat transcript.
A second routine is the “revision receipt”. At the end of a lesson, students highlight three changes they made and label each as one of: clarity, evidence, structure, or voice. Next to each label, they write a one-sentence reason. If a student used rewrite-on-command, they must also note what they asked for and what they rejected. This is where you catch voice drift early, because students are forced to notice when the tone changes.
A third routine is the “source stop”. Before a student is allowed to expand a paragraph, they must add a source note beneath it: where the idea came from, and what would count as evidence. In a history essay, that might be a page number or quotation. In a science explanation, it could be a class practical observation or a textbook diagram. In a personal narrative, it might be sensory detail from a memory map rather than an external citation. The key is that students practise distinguishing between generated phrasing and grounded content.
If you want adaptable lesson moves that work beyond English, the planning patterns in ai-across-the-curriculum-lesson-moves-planning-template can help you build consistent routines so students do not have to relearn expectations in every classroom.
Discover the power of Automated Education by joining out community of educators who are reclaiming their time whilst enriching their classrooms. With our intuitive platform, you can automate administrative tasks, personalise student learning, and engage with your class like never before.
Don’t let administrative tasks overshadow your passion for teaching. Sign up today and transform your educational environment with Automated Education.
🎓 Register for FREE!
Assessment design
Assessment in a co-authoring world improves when the rubric rewards decisions, not just fluency. One approach is to split marks across product and process, with the process marks anchored in the evidence you collected during checkpoints. The final piece still matters, but it is no longer the only artefact that counts.
Task design helps too. Prompts that require situated knowledge reduce generic output. For example, rather than “Write a persuasive letter about school uniforms”, ask students to write a letter to a specific audience using constraints: two class readings, one counterargument drawn from a recent discussion, and a paragraph that anticipates a local concern. In literature, ask for an interpretation that must reference three short quotations, each explained in the student’s own words before any AI rewriting is allowed. In younger year groups, you can do this with oral rehearsal and a simple “tell me your choices” conference, then have students use AI only for sentence-level clarity.
Rubrics can explicitly assess “authorial control”. You might look for evidence that the student can explain their structure, justify a key revision, and defend their source choices. Another criterion can be “integrity of sourcing”, which rewards correct attribution and penalises unsupported claims, regardless of whether AI was used. This reframes the conversation from suspicion to scholarship: show your working, as you would in maths.
When students ask, “Can I use AI?”, your answer becomes, “Yes—if your evidence shows you were the writer making the decisions.” That is a healthier norm than “Yes, but don’t get caught.”
Practical implementation
An evidence-first approach only works if it is manageable. Start with minimum-data rules. Decide what you actually need: perhaps three prompt snapshots per assignment, two revision receipts, and a short source list. Avoid collecting full chat histories by default, especially if privacy policies are unclear. Encourage students to paste only the relevant excerpt into a document they control, and to remove personal data from prompts. If your school uses shared devices, teach students not to enter names, addresses, medical details, or anything that could identify another pupil.
Access and equity need explicit planning. Some students will have better devices, faster internet, or paid features at home. Keep the assessed evidence within lesson time where possible, and offer non-AI pathways that still meet the same learning goals. For instance, a student without access can use peer feedback as their “assistant”, completing the same revision receipt and source stop routines. The assessment should reward the thinking, not the tool.
A one-page student agreement helps set expectations without turning the classroom into a courtroom. Keep it plain and teach it like any other success criteria. It might include: students will keep a prompt log at checkpoints; they will not submit AI-generated sources as real; they will be able to explain their main argument and revisions; they will cite any external material; and they will not use AI to impersonate another person’s voice. You can also include a line about teacher discretion: if the evidence trail is missing, the student may be asked to complete an in-class writing conference or a short, on-demand paragraph to demonstrate understanding.
If you want a ready-made set of activities that build these habits through purposeful writing moments, world-book-day-ai-evidence-pack-classroom-activities offers ideas that translate well beyond the event itself.
Writing instruction has always been about making invisible thinking visible. Co-authoring tools simply raise the stakes. When you collect small, authentic evidence of decisions, students can use powerful tools while still learning to write like authors.
For clearer drafts and more honest revision conversations ahead!
The Automated Education Team