Thoughtful, evidence-informed approaches to AI in education

AI is increasingly present in educational practice, often ahead of shared frameworks, policy clarity, or a consistent understanding of risk. Schools and trusts are navigating complex questions about appropriateness, safeguarding, data protection, and professional judgement.

The Institute of AI in Education (IAIE) exists to support education settings to engage with AI carefully, responsibly, and in line with national expectations. Our work focuses on decision support, professional learning, and structured pilot activity, not on promoting tools or accelerating adoption.

Our approach

IAIE takes a guidance-led, decision-support approach to AI in education.

We start with national guidance, professional standards, and safeguarding expectations, and work collaboratively with educators to explore how AI-related decisions can be made consistently, transparently, and defensibly.

Our work is grounded in three principles:

  • risks must be recognised and managed

  • learning should be evidence-informed

  • professional judgement remains central

Design and Deliver Pilot Programmes

IAIE designs and delivers small-scale, carefully structured pilot programmes with schools, trusts, and partners. These pilots explore how AI can be used appropriately within educational settings, always within clear boundaries of safeguarding, data protection, and professional judgement.

Our pilots are:

  • practitioner-led and context-specific

  • focused on practical classroom and leadership use

  • designed to test feasibility rather than make broad claims

  • time-limited and proportionate in scale

     

Each pilot is treated as a learning opportunity, generating insights that inform future practice and decision-making.

Our Approach in Practice

Across all activity, IAIE is guided by a consistent approach:

Practitioner-led design
Safeguarding and ethics first
Evidence before scale
Transparency and accountability
Respect for the complexity of education

Support Professional Development

We provide professional development for teachers and school leaders focused on responsible AI use in education. Our CPD is designed to be relevant, realistic, and respectful of professional expertise.

Professional development through IAIE:

  • supports understanding of what AI can and cannot do

  • explores ethical, legal, and safeguarding considerations

  • emphasises teacher-in-the-loop decision-making

  • avoids tool-specific or vendor-led training

The aim is not to promote technology adoption, but to build professional confidence and critical understanding.

Promote Responsible and Ethical Practice

Responsible AI use is central to all IAIE activity. We work with schools and partners to consider:

  • safeguarding and pupil protection

  • data privacy and GDPR compliance

  • bias, reliability, and transparency

  • the impact of AI on professional judgement

     

Rather than treating these as add-ons, ethical and safeguarding considerations are embedded from the outset in programme design and delivery.

Generate and Share Evidence

IAIE places strong emphasis on evidence-informed practice. We design our work to generate meaningful insights about:

  • what works in real educational contexts

  • what conditions are needed for responsible use

  • what works in real educational contexts

     

Evaluation is proportionate and practical, avoiding unnecessary burden on educators while still supporting learning and accountability. Findings are used to refine programmes and shared with partners and the wider sector where appropriate.

Work in Partnership with the Sector

We believe sustainable progress comes through collaboration. IAIE works with:

  • schools and multi-academy trusts

  • researchers and policy-aligned partners

  • funders and public-interest organisations

  • researchers and policy-aligned partner

     

Have a question for IAIE?