4 'Key Foundations' for AI in Education

Artificial intelligence (AI) is here to stay. The U.S Department of Education has released guidance for the education sector — including potential benefits, warnings about misuses, and principles for developing guidelines at the district level.

During hearings in the U.S. Congress in May 2023, AI developers and key stakeholders urged lawmakers to take seriously the need for oversight of this technology and to act now to develop policies and laws to monitor and direct the responsible use of it. Every sector is being affected by AI, including the education system.

In its report, “Artificial Intelligence and the Future of Teaching and Learning: Insights and Recommendations,” released in May 2023, the Office of Educational Technology (OET) of the U.S. Department of Education (ED) outlines several core messages. ED makes clear its support for ed tech, including AI, to improve teaching and learning but emphasizes that knowledge about AI needs to be shared, support given for those using it, and policies developed for its safe and responsible use.

The report recognizes that the main components of AI are being “increasingly embedded in all types of educational systems,” shifting from “capturing data to detecting patterns in data” and from “providing access to instructional resources to automating decisions.”

It expresses deep concerns about the increase in delegating responsibilities to a computer system. The report also expresses concerns about biases in pattern detection and fairness in automating decisions.

The report acknowledges the opportunities AI presents: voice assistants, writing help, trip planning, and, for educators, disabled student support, multilingual learning support, finding lesson materials, and lesson creation support.

But there are risks as well: security and privacy breaches, inappropriate or wrong information, enhancement of cultural or other biases, student plagiarism, and fairness.

The report also acknowledges the rapid rise and the benefits and risks of generative AI-enabled chatbots, but does not focus specifically on those.

During listening sessions in 2022 and 2023, attended by over 700 participants, OET recognized three main reasons to address the use of AI in education now: First, AI can help educate students at a lower cost, help them make up learning losses due to the pandemic, and customize learning to their specific culture and community.

Second, risks include increased student surveillance, dangers of “algorithmic discrimination” against some student populations, unfair exam monitoring systems, or inaccurate information.

Third, unintended or unexpected consequences may include the AI adapting its curricular learning pace for specific students based on “incomplete data, poor theories, or biased assumptions about learning,” thus affecting achievement rates. This can also affect the hiring of teachers.

With these issues in mind, the report articulates guiding questions about what a collective vision for the use of AI in education looks like and what the timeline should be for developing guidelines to address it, based on four “key foundations”:

  1. Which AI technologies to use in education to keep people central in decision-making;
  2. How to advance equity and root out bias that could interfere with every student achieving success;
  3. How to ensure safety, ethics, and effectiveness in data privacy and security;
  4. How to ensure transparency and disclosure about how AI models work in terms of educational goals, as well as human control when needed.

Visit this page to read and download a summary handout of the report’s main points. A webinar going into more depth on this report will be held Tuesday, June 13, 2023, at 2:30 p.m. ET. Signup is available by QR code at this summary handout link.

The full report can be downloaded from this page.

About the Author

Kate Lucariello is a former newspaper editor, EAST Lab high school teacher and college English teacher.

Featured

  • abstract pattern of cybersecurity, ai and cloud imagery

    Report Identifies Malicious Use of AI in Cloud-Based Cyber Threats

    A recent report from OpenAI identifies the misuse of artificial intelligence in cybercrime, social engineering, and influence operations, particularly those targeting or operating through cloud infrastructure. In "Disrupting Malicious Uses of AI: June 2025," the company outlines how threat actors are weaponizing large language models for malicious ends — and how OpenAI is pushing back.

  • glowing shield hovers above a digital cloud platform with abstract data streams and cloud icons in the background

    Google to Acquire Cloud Security Firm Wiz in $32 Billion Deal

    Google has announced it will acquire cloud security startup Wiz for $32 billion. If completed, the acquisition — an all-cash deal — would mark the largest in Google's history.

  • laptop and fish hook

    Security Researchers Identify Generative AI 'Vishing' Attack

    A new report from researchers at Ontinue's Cyber Defense Center has identified a complex, multi-stage cyber attack that leveraged social engineering, remote access tools, and signed binaries to infiltrate and persist within a target network.

  • glowing futuristic laptop with a holographic screen displaying digital text

    New Turnitin Product Offers AI-Powered Writing Tools with Instructor Guardrails

    Academic integrity solution provider Turnitin has launched Turnitin Clarity, a paid add-on for Turnitin Feedback Studio that provides a composition workspace for students with educator-guided AI assistance, AI-generated writing feedback, visibility into integrity insights, and more.