ChatGPT: Consider reviewing homework polices, DfE tells colleges

The government has set out its stance on the use of generative AI in education

The government has set out its stance on the use of generative AI in education

Schools and colleges “may wish to review” their homework policies amid fears about the use of artificial intelligence like ChatGPT, the Department for Education has warned. 

The department has set out its stance on generative AI in a statement published this morning.

Exam boards yesterday published their own guidance on “protecting the integrity of qualifications” from AI.

The DfE said that when used “appropriately” it has potential to reduce workload across the education sector and free up teachers’ time. 

But said schools and colleges “may wish to review homework policies, to consider the approach to homework and other forms of unsupervised study as necessary to account for the availability of generative AI”. 

It follows reports of schools abandoning homework essays because of AI.

Schools and colleges should also “review and strengthen” their cyber security as AI could “increase the sophistication and credibility of attacks”. 

Students should be protected from harmful online content and personal, sensitive data should not be entered into AI tools, DfE said. 

The department warned the quality and content of any final documents – such as administrative plans – remains the “professional responsibilities of the person who produces it and the organisation they belong to.” 

Education sector has ‘lagged in tech adoption’

The DfE will now convene experts to work with the education sector and “share and identify best practice and opportunities to improve education and reduce workload using generative AI”. 

They say to “harness the potential” of AI, students will need to be “knowledgeable and develop their intellectual capability”.

“The education system should support students, particularly young pupils, to identify and use appropriate resources to support their ongoing education. 

“This includes encouraging effective use of age-appropriate resources (which in some instances may include generative AI) and preventing over-reliance on a limited number of tools or resources.”

Speaking at Bett Show ed tech conference this morning, education secretary Gillian Keegan said the education sector has “often lagged in tech adoption” and is a tool schools and colleges “haven’t yet managed to get the most out of”.

She said that tech that doesn’t work is an “expensive and potentially dangerous mistake” and one that education providers “cannot afford to make”. 

Keegan believes teachers’ work could be “transformed” by AI but it’s not yet at the standard needed. 

Sector ‘moving too slow’ on AI

Education experts were quizzed by MPs on using AI in education at the science and technology committee this morning. 

Rose Luckin, professor of learned centred design at University College London, warned that the education sector doesn’t have “the in-depth knowledge about AI to be able to do a really good job.

“The technology’s moving at pace, it’s increasingly complex. Even the people developing it don’t always understand the implications of what it does. 

Daisy Christodoulou, director of education at No More Marking, said “speed matters” when responding to AI changes and that too many organsiations are “moving very, very slowly”.

“I think we need to have a good, hard look at how we assess. I do think ChatGPT has huge implications for continuous assessment and coursework,” she said. 

“I’ve heard a few suggestions about different things you could do…but some of the people making those suggestions don’t realise quite how powerful a tool like ChatGPT is.

It is capable of producing, original, very hard to detect, relatively high-quality responses to any kind of question. We have to be looking at assessments that are in more controlled environments.”

Latest education roles from

Head of Safeguarding & Wellbeing

Head of Safeguarding & Wellbeing

Capital City College Group

Group Principal & Chief Executive Officer

Group Principal & Chief Executive Officer

Windsor Forest Colleges Group

Regional Director

Regional Director

Leo Academy Trust

Executive Head Teacher (Trust-wide SEND)

Executive Head Teacher (Trust-wide SEND)

The Legacy Learning Trust

Sponsored posts

Sponsored post

Preparing learners for work, not just exams: the case for skills-led learning

As further education (FE) continues to adapt to shifting labour markets, digital transformation and widening participation agendas, providers are...

Advertorial
Sponsored post

How Eduqas GCSE English Language is turning the page on ‘I’m never going to pass’

“A lot of learners come to us thinking ‘I’m rubbish at English, and I’m never going to pass’,” says...

Advertorial
Sponsored post

Fragmentation in FE: tackling the problem of disjointed tech, with OneAdvanced Education

Further education has always been a place where people make complexity work through dedication and ingenuity. Colleges and apprenticeship...

Advertorial
Sponsored post

Teaching leadership early: the missing piece in youth employability

Leaders in education and industry are ready to play their part in tackling the UK’s alarming levels of youth...

Advertorial

More from this theme

AI, Skills reform

AI Skills Hub risks ‘copy and paste of past failure’

New AI skills hub initiative reeks of pandemic-era 'skills toolkits' failures

Anviksha Patel
AI

Ofsted reveals how it will inspect providers’ AI use

Inspectors will not check tech use as a ‘standalone’ part of inspections, but will look at its impact on...

Jack Dyson
AI, Colleges

AI guidance for colleges: 9 key findings for leaders

Government toolkits say colleges should train staff on safe AI use and to spot deep-fakes

Jack Dyson
AI

FE providers wanted to become edtech ‘testbeds’

Pilot to build 'evidence base' on impact of workload-cutting tech

Jack Dyson

Your thoughts

Leave a Reply

Your email address will not be published. Required fields are marked *

One comment

  1. Is it out of the realm of possibility for (eventually) to use AI in order to detect cheating? Such as AI can analyse and compare work students do supervised in class, and compare it to work they do at home on their own and submit?

    If AI is smart enough to write the bulk of an essay surely it can be used to compare and contrast material to spot inconsistencies.