What GPT-5 Might Mean for Schools

A procurement-and-practice stress test for leaders

A school leadership team reviewing AI readiness plans on a laptop

What we mean by “GPT-5”

In this article, “GPT-5” is shorthand for the next noticeable jump in general-purpose AI models used in education: systems that can read, write, listen and look across longer stretches of information, and then act with more confidence. That might be a new version from one provider, or a set of competing models that reach a similar level within months of each other. Either way, the practical question for schools is not “Which model wins?” but “Which processes break, bend or improve when AI becomes more capable?”

It is also worth stating what we do not mean. We are not assuming a flawless, human-level tutor that never hallucinates, never mishandles data, and never amplifies bias. We are not assuming that every classroom will immediately adopt AI-heavy workflows. And we are not assuming that the best approach is to buy the biggest tool and hope for the best. Instead, treat “GPT-5” as a procurement-and-practice stress test: if capability rises, where are your current policies vague, where are your routines fragile, and where are your staff unsupported?

If your team is tracking the wider landscape, it helps to keep a running “what changed?” log rather than reacting to headlines. You might find it useful to pair this with your existing horizon scanning, such as an annual tools review like AI tools refresh 2025, so you can separate genuine capability shifts from marketing noise.

Assumptions and uncertainty

Planning for uncertainty is a leadership skill, not a technical one. A sensible approach is to define a small set of plausible “capability jumps” and test your school against each. If the jump arrives, you are ready; if it arrives later, the work still improves safeguarding, data handling and teaching clarity.

A practical way to do this is to write down three to five assumptions you are willing to plan around for the next two terms, and three you are not willing to assume. For example, you might plan around “AI can handle longer documents and more images reliably” while refusing to assume “AI outputs are accurate enough to publish without checking”. This keeps your policy grounded and your staff confident.

Likely capability shifts

The changes that matter most in schools are rarely “new features”. They are shifts in reliability, reach and integration. Longer context means the AI can take a whole scheme of learning, a set of assessment papers and a department policy, and respond in a way that appears coherent. Stronger reasoning means it can plan multi-step tasks and justify choices more convincingly, which can be helpful for modelling—but risky if staff over-trust it. More reliable multimodal understanding means it can interpret a photo of a science practical set-up, a scan of student work, or a diagram from a textbook with fewer errors. Early-stage agentic actions mean it can do things on your behalf: draft emails, move files, populate a template, or trigger actions across connected systems.

These shifts do not just change what teachers can do. They change what teachers might be expected to do, and what leaders might assume is “efficient”. That is why your readiness plan must protect professional judgement and workload, not just add another tool.

Scenario 1: More multimodal use

As multimodal AI becomes more dependable, classroom use will move beyond “write me a paragraph” into “look at this and help me improve it”. In English, that might be a student photographing an annotated poem and asking for a clearer explanation of imagery, then comparing the AI’s interpretation with the class discussion. In maths, it might be a learner uploading a worked solution and asking where their reasoning went off track. In art and design, it could be a critique of composition choices, with the teacher guiding students to articulate intent and evaluate suggestions rather than accept them.

Practical subjects and accessibility are where the benefits and risks both sharpen. A technology teacher might use AI to generate step-by-step safety reminders for a workshop routine, tailored to the specific equipment pictured. A PE teacher might use it to create alternative drills for an injured student after describing the lesson context. For accessibility, multimodal AI can turn dense worksheets into simplified versions, generate captions for short clips, or describe images for learners with visual impairments—provided you have clear rules about what data is shared and how outputs are checked.

The disruption here is not just pedagogical. It is procedural. If students can submit photos of work and get instant feedback, your homework policy, feedback expectations and academic integrity guidance will need updating. You will also need a clear stance on whether student images, voices and work samples can be processed by third-party systems, and under what conditions.

Scenario 2: End-to-end workflows

The next shift is “end-to-end” workflows: planning, assessment and feedback stitched together. An AI system with longer context might take a unit plan, generate a draft assessment, propose a rubric aligned to your success criteria, create differentiated versions, and then draft feedback comments after reading student responses. That can save time, but it can also quietly standardise your curriculum in ways you did not choose.

This is where humans must stay firmly in the loop. Rubrics encode values. Feedback shapes motivation. Moderation protects fairness. If AI proposes a rubric, the department still decides what excellence looks like. If AI drafts feedback, the teacher still checks tone, accuracy and next steps. If AI suggests grade boundaries, leaders still validate against evidence and professional judgement.

A useful mental model is: let AI do the first draft and the tedious formatting; keep humans responsible for the decisions that affect students’ outcomes and wellbeing. If you want to explore how “computer-use” style assistants might influence these workflows, the scenario planning in Claude computer use: school systems assistant can help leaders see where convenience becomes risk.

Ready to Revolutionise Your Teaching Experience?

Discover the power of Automated Education by joining out community of educators who are reclaiming their time whilst enriching their classrooms. With our intuitive platform, you can automate administrative tasks, personalise student learning, and engage with your class like never before.

Don’t let administrative tasks overshadow your passion for teaching. Sign up today and transform your educational environment with Automated Education.

🎓 Register for FREE!

Scenario 3: Early agentic behaviour

Agentic behaviour is where the risk surface changes. When AI can take actions across files, email and connected platforms, mistakes stop being confined to a chat window. A well-meaning prompt like “Send parents a reminder about tomorrow’s trip” becomes high-stakes if the AI selects the wrong contact group, includes sensitive details, or sends without approval. Similarly, “Update the seating plan and note behaviour concerns” becomes problematic if it writes subjective labels into a system of record.

Even early-stage agents tend to fail in predictable ways: they misunderstand ambiguous instructions, overreach their permissions, and behave inconsistently across edge cases. Schools should assume that agentic tools will arrive before governance is mature. That means you need to define permission boundaries now: what can be read, what can be written, what requires human approval, and what must never be automated.

This is also where integration with MIS/LMS matters. If an AI tool can “see” attendance, behaviour logs or safeguarding notes, your data protection and safeguarding posture must be explicit. Treat every integration as a new member of staff: it needs role-based access, audit trails, and clear accountability.

The risk register

A practical risk register for “GPT-5 level” capability should cover safeguarding, privacy, bias, reliability, academic integrity, copyright, vendor lock-in and cost volatility. Safeguarding includes both content risk (unsafe advice, grooming-like interactions, inappropriate images) and contact risk (who can message whom, and through what channel). Privacy includes what data is processed, where it goes, how long it is stored, and whether it is used to train models. Bias remains a concern in feedback tone, behavioural interpretations and assumptions about students’ backgrounds.

Reliability is not just about factual errors. It is about consistency: will the same prompt produce materially different guidance on different days, and can staff recognise when it is “confidently wrong”? Academic integrity shifts as AI becomes better at interpreting prompts and generating plausible student-like work, including multimodal artefacts. Copyright is increasingly relevant when AI summarises textbooks, interprets images, or generates “in the style of” resources. Vendor lock-in and cost volatility matter because pricing and access can change quickly; leaders need exit plans and budget resilience.

If you want a model-specific example of how quickly perceptions can change, the discussion in DeepSeek R1 schools briefing is a reminder that capability, cost and governance can move in unexpected directions.

Procurement readiness

Procurement questions should shift from “What can it do?” to “What controls and evidence exist?” Ask vendors to show, not tell. You are looking for clear documentation and practical demonstrations of data protection, logging, retention, model routing and admin controls.

At minimum, ask how data is processed and stored, what is retained by default, and what you can switch off. Ask whether prompts and uploads are used for training, and what contractual terms guarantee this. Ask what logs are available to administrators, how long they are kept, and whether you can export them for audit. Ask how the tool routes requests between models, especially if it uses multiple providers behind the scenes; “model routing” can introduce hidden data flows and inconsistent behaviour. Ask what role-based access controls exist, whether you can restrict features by age group, and whether you can disable multimodal uploads or external integrations.

Evidence matters. Request a data protection pack, a clear retention schedule, and a walkthrough of the admin console. Ask for examples of safety mitigations, including how the system handles self-harm content, sexual content and personal data. If the tool offers agentic actions, require an approval step and an audit trail, and insist on a permissions model that matches school roles.

Policy and governance updates

When capability jumps, policies often become either too strict to follow or too vague to protect anyone. The goal is to tighten what matters and simplify what staff need day to day. Tighten rules around personal data, student images and voice, system integrations, and any automation that writes to official records. Simplify classroom guidance into routines teachers can remember: what’s allowed, what needs permission, what must be checked, and what must never be entered.

Communication matters as much as the policy text. Staff need clarity on what “good use” looks like, students need explicit boundaries and rationale, and parents need reassurance about data and safeguarding. If your leadership team is monitoring regulatory shifts, keep a living addendum rather than rewriting the whole policy each time; updates like those discussed in AI policy watch: government updates can be translated into plain-language changes for staff.

Staff training

A minimum viable set of competencies will serve you better than one-off “AI twilight” sessions. Staff need to understand how to prompt for their context, how to check outputs, and how to spot common failure modes. They also need practical safeguarding habits: never entering sensitive data, using anonymised examples, and escalating concerns when AI outputs are harmful or confusing. Finally, they need a shared language for academic integrity so that expectations are consistent across subjects.

Build CPD that survives term time by embedding it into existing structures. A short department discussion during planning meetings can focus on one routine, such as “AI-assisted feedback: what we allow and how we check tone and accuracy”. A pastoral briefing can cover “AI and safeguarding: what to do if a student discloses something to a chatbot”. A learning walk focus can include “students explaining how they used AI”, which reinforces metacognition rather than secrecy.

Readiness checklist and plan

A readiness checklist works best when it is owned across roles. Leaders should confirm the school’s risk appetite and non-negotiables, and ensure there is a named owner for AI governance. IT should map data flows, integrations and permissions, and confirm logging and incident response. The DSL should stress-test safeguarding scenarios, including how AI tools handle disclosures and harmful content. Heads of department should define acceptable classroom routines and assessment boundaries. Classroom teachers should trial low-risk use cases and feed back what actually happens with real students.

In the next 30 days, focus on visibility and boundaries. Audit which AI tools are already in use, including “free” accounts staff and students may be using. Publish a one-page interim guidance note that clarifies data rules, acceptable use and escalation routes. Start a vendor question bank so procurement is consistent, and identify one or two low-risk pilot routines, such as lesson resource drafting with no student data.

In 60 days, move to controlled practice. Run short CPD cycles in departments, agree subject-specific integrity expectations, and test your incident process with a tabletop exercise: for example, an AI tool generating unsafe advice, or an agent sending an email to the wrong group. Begin procurement conversations using your evidence requirements, and ensure your admin controls, logging and retention settings are understood and documented.

In 90 days, aim for sustainable governance. Finalise policy updates, publish student and parent guidance, and establish a review cadence for tools and risks. If you pilot an integrated tool, ensure role-based access, approval steps for actions, and audit trails are in place before wider rollout. Most importantly, evaluate impact on workload and learning, not just novelty. If a workflow saves time but reduces feedback quality, it is not a win.

Preparing for “GPT-5” is ultimately about strengthening the muscles schools need for any fast-moving technology: clear boundaries, good procurement questions, realistic training, and routines that protect students while supporting teachers.

To calmer rollouts and clearer routines ahead, The Automated Education Team

Table of Contents

Categories

Education

Tags

AI in Education Technology Teacher Training

Latest

Alternative Languages