Site icon Al Kingsley MBE – Education, Technology, Leadership #EdTech #Ai

PISA 2029 “Media & Artificial Intelligence Literacy (MAIL)”

1) What MAIL is (and what it isn’t)

MAIL is PISA 2029’s “innovative domain”: a cross-cutting assessment looking at how well 15-year-olds can engage with digital media and AI-mediated content effectively, ethically and responsibly.

It’s not a test of “how to use apps”, coding, or who’s best at prompting. It’s closer to:

2) Why it matters (strategically, not just academically)

MAIL is a signal that international systems are now treating information judgement + AI understanding as core competencies for:

For Trusts, it’s also a governance issue: if students are using AI and social platforms daily, the question becomes whether the school has made them more capable or simply more exposed.

3) What the OECD says it will assess (the likely competency “shape”)

The OECD describes MAIL as the competencies required to interact with digital/AI content effectively, ethically, and responsibly. In practice, this typically implies students can:

4) How it’s expected to be assessed (and why that changes teaching)

MAIL is planned to run in simulated environments that may resemble:

That matters because it pushes beyond “write an essay about misinformation” into performance tasks: making decisions, checking sources, weighing trade-offs, and explaining reasoning.

5) Timeline and what we know right now

6) The related AI Literacy Framework you should track (useful now)

A major feeder into MAIL is the EU–OECD AI Literacy Framework (review draft) (“Empowering Learners for the Age of AI”), supported by Code.org and an expert group. It explicitly states it contributes to the PISA 2029 MAIL assessment.

In plain terms, it frames AI literacy around:

 

What Trusts should do in 2026–27

A) Set “MAIL readiness” as a governance line of sight

Add a standing line to committee oversight (curriculum/standards + safeguarding + digital strategy):

B) Map current provision quickly (you’ll find gaps)

Do a lightweight audit against three buckets:

  1. Credibility & evidence (source checking, lateral reading, recognising persuasion)
  2. Platform & AI awareness (feeds, ranking, deepfakes, generative AI limits)
  3. Ethics & safety (privacy, consent, bias, manipulation, data trails)

Output: a one-page heatmap by key stage/subject.

C) Build “authentic tasks” into existing subjects

Examples (low workload, high signal):

D) Sort staff confidence (CPD that doesn’t insult professionals)

Prioritise three staff competencies:

E) Align policy: safeguarding + assessment + acceptable use

Governors should expect coherence across:

F) Measure progress simply (don’t wait for PISA)

Use internal indicators twice yearly:

Intentionally simple Questions governors could ask (useful in minutes)

  1. If PISA dropped a simulated “AI + social feed” task into Year 10 tomorrow, how would our pupils cope?
  2. Where is this taught explicitly (not assumed) and where is it practised?
  3. Are we building capability… or just writing policies?
  4. What’s our stance on AI use for learning vs. simple time-saving shortcuts, and do students understand it?
  5. How do we know disadvantaged students aren’t being left to “figure this out” alone?

 

Bottom line

MAIL is essentially the OECD saying: critical literacy now includes AI and platform reality. Treat it as a curriculum priority and a governance priority, and you’ll be ahead of the curve well before 2029.

Download a hi-res copy of both infographics from here (1) and (2) in PNG format.

Exit mobile version