Education
How to navigate the future of AI in education and education in AI | EY
Policymakers and education leaders play a crucial role in mitigating risks and challenges associated with AI in education. They can achieve this by implementing robust policy frameworks. While adopting AI-driven digital tools has the potential to widen the educational divide, it also introduces various risks, including impacts on academic integrity, with concerns about plagiarism, cheating in assessments, and the loss of critical skills. Additionally, there are risks related to inaccuracy, discrimination, bias and data privacy.
Teachers will remain essential in education, even with AI-powered tutors. AI lacks the nuanced understanding and empathy of human educators, making it less effective in addressing complex learning needs. Additionally, AI tools have limitations in content recency and accuracy, often “hallucinating” or fabricating information. Students and teachers must learn to navigate these limitations and fact-check AI outputs. Consequently, AI tutors cannot yet be trusted to educate young children who may not recognize errors or hallucinations.
Transparency and explainability
AI models can function as “black boxes,” with their inner workings often hidden or poorly understood. This lack of transparency complicates users’ ability to understand the rationale behind decisions related to learning pathways, grading, or student performance, potentially introducing bias or errors. Without transparency and explainability, students, parents, and teachers cannot assess the fairness or accuracy of AI-based recommendations.
To promote widespread acceptance and effective use of AI in education, it’s essential to establish protocols for model development that clarify how AI recommendations are generated and validated. Building transparency and explainability into AI models is vital for fostering confidence.
Privacy and data security
Privacy and data security are foundational for building confidence in AI-powered educational tools, promoting technology that enhances learning without compromising the safety and rights of students. In education, AI tools often rely on large amounts of student data to provide personalized learning experiences, track progress or adapt to individual needs However, collecting and storing sensitive information, such as academic performance and behavioral patterns, raises concerns about who has access to this data and how it can be used.
Impact on quality of learning and academic integrity
Educators are concerned that over-reliance on AI tools, particularly GenAI, may undermine students’ skills in research, critical thinking, and effective communication. GenAI raises academic integrity issues, as it can produce essays, solve math problems, and complete tests that seem acceptable. Consequently, the focus has often been on prohibiting AI use or detecting AI-generated submissions. However, students need clear guidance and policies on the ethical and effective use of AI.
As schools and universities adopt AI-driven approaches, they should regularly monitor and evaluate outcomes while gathering stakeholder feedback, especially regarding equity. This enables necessary adjustments to enhance effectiveness and address any gaps.
Lack of digital and AI skills among staff and students
Most teachers have not yet integrated AI into their daily workflows, primarily due to a lack of knowledge on how to do so. They urgently need guidance, support, and resources to effectively use AI tools to enhance learning. A 2024 report by EdWeek Research Center found that the primary reason US teachers have not incorporated AI into their practice is a “lack of knowledge and support”.
Lack of digital infrastructure
To leverage AI in education, teachers and learners need access to digital devices, high-speed internet, and secure platforms. Without universal technology access, the digital divide and educational inequity will worsen, particularly in underserved communities lacking reliable internet and resources for computers. Promoting equitable AI-enhanced education requires investment in high-speed internet, connected devices, and, in some regions, stable electricity.
AI educational tools like adaptive learning platforms and intelligent tutoring systems require significant financial investment for setup, maintenance, updates, and data security. Implementing AI necessitates sophisticated data systems for managing student data, which must adhere to strict privacy standards, increasing costs for software, data protection, and staff training. Many schools, especially in underserved areas, struggle to fund traditional resources, let alone advanced technologies and their security needs.
Education
9 AI Ethics Scenarios (and What School Librarians Would Do)
A common refrain about artificial intelligence in education is that it’s a research tool, and as such, some school librarians are acquiring firsthand experience with its uses and controversies.
Leading a presentation last week at the ISTELive 25 + ASCD annual conference in San Antonio, a trio of librarians parsed appropriate and inappropriate uses of AI in a series of hypothetical scenarios. They broadly recommended that schools have, and clearly articulate, official policies governing AI use and be cautious about inputting copyrighted or private information.
Amanda Hunt, a librarian at Oak Run Middle School in Texas, said their presentation would focus on scenarios because librarians are experiencing so many.
“The reason we did it this way is because these scenarios are coming up,” she said. “Every day I’m hearing some other type of question in regards to AI and how we’re using it in the classroom or in the library.”
- Scenario 1: A class encourages students to use generative AI for brainstorming, outlining and summarizing articles.
Elissa Malespina, a teacher librarian at Science Park High School in New Jersey, said she felt this was a valid use, as she has found AI to be helpful for high schoolers who are prone to get overwhelmed by research projects.
Ashley Cooksey, an assistant professor and school library program director at Arkansas Tech University, disagreed slightly: While she appreciates AI’s ability to outline and brainstorm, she said, she would discourage her students from using it to synthesize summaries.
“Point one on that is that you’re not using your synthesis and digging deep and reading the article for yourself to pull out the information pertinent to you,” she said. “Point No. 2 — I publish, I write. If you’re in higher ed, you do that. I don’t want someone to put my work into a piece of generative AI and an [LLM] that is then going to use work I worked very, very hard on to train its language learning model.”
- Scenario 2: A school district buys an AI tool that generates student book reviews for a library website, which saves time and promotes titles but misses key themes or introduces unintended bias.
All three speakers said this use of AI could certainly be helpful to librarians, but if the reviews are labeled in a way that makes it sound like they were written by students when they weren’t, that wouldn’t be ethical.
- Scenario 3: An administrator asks a librarian to use AI to generate new curriculum materials and library signage. Do the outputs violate copyright or proper attribution rules?
Hunt said the answer depends on local and district regulations, but she recommended using Adobe Express because it doesn’t pull from the Internet.
- Scenario 4: An ed-tech vendor pitches a school library on an AI tool that analyzes circulation data and automatically recommends titles to purchase. It learns from the school’s preferences but often excludes lesser-known topics or authors of certain backgrounds.
Hunt, Malespina and Cooksey agreed that this would be problematic, especially because entering circulation data could include personally identifiable information, which should never be entered into an AI.
- Scenario 5: At a school that doesn’t have a clear AI policy, a student uses AI to summarize a research article and gets accused of plagiarism. Who is responsible, and what is the librarian’s role?
The speakers as well as polled audience members tended to agree the school district would be responsible in this scenario. Without a policy in place, the school will have a harder time establishing whether a student’s behavior constitutes plagiarism.
Cooksey emphasized the need for ongoing professional development, and Hunt said any districts that don’t have an official AI policy need steady pressure until they draft one.
“I am the squeaky wheel right now in my district, and I’m going to continue to be annoying about it, but I feel like we need to have something in place,” Hunt said.
- Scenario 6: Attempting to cause trouble, a student creates a deepfake of a teacher acting inappropriately. Administrators struggle to respond, they have no specific policy in place, and trust is shaken.
Again, the speakers said this is one more example to illustrate the importance of AI policies as well as AI literacy.
“We’re getting to this point where we need to be questioning so much of what we see, hear and read,” Hunt said.
- Scenario 7: A pilot program uses AI to provide instant feedback on student essays, but English language learners consistently get lower scores, leading teachers to worry the AI system can’t recognize code-switching or cultural context.
In response to this situation, Hunt said it’s important to know whether the parent has given their permission to enter student essays into an AI, and the teacher or librarian should still be reading the essays themselves.
Malespina and Cooksey both cautioned against relying on AI plagiarism detection tools.
“None of these tools can do a good enough job, and they are biased toward [English language learners],” Malespina said.
- Scenario 8: A school-approved AI system flags students who haven’t checked out any books recently, tracks their reading speed and completion patterns, and recommends interventions.
Malespina said she doesn’t want an AI tool tracking students in that much detail, and Cooksey pointed out that reading speed and completion patterns aren’t reliably indicative of anything that teachers need to know about students.
- Scenario 9: An AI tool translates texts, reads books aloud and simplifies complex texts for students with individualized education programs, but it doesn’t always translate nuance or tone.
Hunt said she sees benefit in this kind of application for students who need extra support, but she said the loss of tone could be an issue, and it raises questions about infringing on audiobook copyright laws.
Cooksey expounded upon that.
“Additionally, copyright goes beyond the printed work. … That copyright owner also owns the presentation rights, the audio rights and anything like that,” she said. “So if they’re putting something into a generative AI tool that reads the PDF, that is technically a violation of copyright in that moment, because there are available tools for audio versions of books for this reason, and they’re widely available. Sora is great, and it’s free for educators. … But when you’re talking about taking something that belongs to someone else and generating a brand-new copied product of that, that’s not fair use.”
Education
Bret Harte Superintendent Named To State Boards On School Finance And AI
Education
Blunkett urges ministers to use ‘incredible sensitivity’ in changing Send system in England | Special educational needs
Ministers must use “incredible sensitivity” in making changes to the special educational needs system, former education secretary David Blunkett has said, as the government is urged not to drop education, health and care plans (EHCPs).
Lord Blunkett, who went through the special needs system when attending a residential school for blind children, said ministers would have to tread carefully.
The former home secretary in Tony Blair’s government also urged the government to reassure parents that it was looking for “a meaningful replacement” for EHCPs, which guarantee more than 600,000 children and young people individual support in learning.
Blunkett said he sympathised with the challenge facing Bridget Phillipson, the education secretary, saying: “It’s absolutely clear that the government will need to do this with incredible sensitivity and with a recognition it’s going to be a bumpy road.”
He said government proposals due in the autumn to reexamine Send provision in England were not the same as welfare changes, largely abandoned last week, which were aimed at reducing spending. “They put another billion in [to Send provision] and nobody noticed,” Blunkett said, adding: “We’ve got to reduce the fear of change.”
Earlier Helen Hayes, the Labour MP who chairs the cross-party Commons education select committee, called for Downing Street to commit to EHCPs, saying this was the only way to combat mistrust among many families with Send children.
“I think at this stage that would be the right thing to do,” she told BBC Radio 4’s Today programme. “We have been looking, as the education select committee, at the Send system for the last several months. We have heard extensive evidence from parents, from organisations that represent parents, from professionals and from others who are deeply involved in the system, which is failing so many children and families at the moment.
“One of the consequences of that failure is that parents really have so little trust and confidence in the Send system at the moment. And the government should take that very seriously as it charts a way forward for reform.”
A letter to the Guardian on Monday, signed by dozens of special needs and disability charities and campaigners, warned against government changes to the Send system that would restrict or abolish EHCPs.
Labour MPs who spoke to the Guardian are worried ministers are unable to explain essential details of the special educational needs shake-up being considered in the schools white paper to be published in October.
Downing Street has refused to rule out ending EHCPs, while stressing that no decisions have yet been taken ahead of a white paper on Send provision to be published in October.
Keir Starmer’s deputy spokesperson said: “I’ll just go back to the broader point that the system is not working and is in desperate need of reform. That’s why we want to actively work with parents, families, parliamentarians to make sure we get this right.”
after newsletter promotion
Speaking later in the Commons, Phillipson said there was “no responsibility I take more seriously” than that to more vulnerable children. She said it was a “serious and complex area” that “we as a government are determined to get right”.
The education secretary said: “There will always be a legal right to the additional support children with Send need, and we will protect it. But alongside that, there will be a better system with strengthened support, improved access and more funding.”
Dr Will Shield, an educational psychologist from the University of Exeter, said rumoured proposals that limit EHCPs – potentially to pupils in special schools – were “deeply problematic”.
Shield said: “Mainstream schools frequently rely on EHCPs to access the funding and oversight needed to support children effectively. Without a clear, well-resourced alternative, families will fear their children are not able to access the support they need to achieve and thrive.”
Paul Whiteman, general secretary of the National Association of Head Teachers, said: “Any reforms in this space will likely provoke strong reactions and it will be crucial that the government works closely with both parents and schools every step of the way.”
-
Funding & Business1 week ago
Kayak and Expedia race to build AI travel agents that turn social posts into itineraries
-
Jobs & Careers7 days ago
Mumbai-based Perplexity Alternative Has 60k+ Users Without Funding
-
Mergers & Acquisitions7 days ago
Donald Trump suggests US government review subsidies to Elon Musk’s companies
-
Funding & Business7 days ago
Rethinking Venture Capital’s Talent Pipeline
-
Jobs & Careers6 days ago
Why Agentic AI Isn’t Pure Hype (And What Skeptics Aren’t Seeing Yet)
-
Funding & Business4 days ago
Sakana AI’s TreeQuest: Deploy multi-model teams that outperform individual LLMs by 30%
-
Jobs & Careers7 days ago
Astrophel Aerospace Raises ₹6.84 Crore to Build Reusable Launch Vehicle
-
Jobs & Careers7 days ago
Telangana Launches TGDeX—India’s First State‑Led AI Public Infrastructure
-
Funding & Business1 week ago
From chatbots to collaborators: How AI agents are reshaping enterprise work
-
Jobs & Careers4 days ago
Ilya Sutskever Takes Over as CEO of Safe Superintelligence After Daniel Gross’s Exit