Close Menu
Creative Learning GuildCreative Learning Guild
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    Creative Learning GuildCreative Learning Guild
    Subscribe
    • Home
    • All
    • News
    • Trending
    • Celebrities
    • Privacy Policy
    • About
    • Contact Us
    • Terms Of Service
    Creative Learning GuildCreative Learning Guild
    Home » The Third-Grade Experiment: What Happened When Children Were Asked to Govern Their Own AI Rules
    AI

    The Third-Grade Experiment: What Happened When Children Were Asked to Govern Their Own AI Rules

    Errica JensenBy Errica JensenApril 23, 2026No Comments5 Mins Read
    Facebook Twitter Pinterest LinkedIn Tumblr Email
    Share
    Facebook Twitter LinkedIn Pinterest Email

    A group of eight and nine-year-olds sat around tables covered in large sheets of paper, markers in hand, on a Tuesday morning in April in a classroom on the third floor of DREAM Charter School in East Harlem. They were working on a task that the New York City Department of Education had yet to complete: writing the rules for artificial intelligence.

    The hand-drawn borders, block lettering, and sporadic misspellings on the posters they created represented concepts that policy staff and education researchers had been discussing for months in conference rooms. “Use your brain first.” “Don’t copy and paste.” “Check your work.” “AI can turn our brains into mush,” a third-grader named David Ortiz said, summarizing the main issue with a directness that no adult in any policy document has quite matched. “It would be no point of school if AI is going to tell you everything.”

    An eight-year-old’s statement is impressive. The phenomenon that cognitive scientists refer to as “cognitive offloading”—the process by which machine reasoning gradually replaces human reasoning, leaving the human less skilled, less capable, and less conscious of the loss—was what made David’s discovery remarkable, not because kids are gullible and occasionally surprising. In a recent study, MIT researchers asked three groups to write essays: one group used ChatGPT, one group used Google, and one group used only their minds. With every assignment, the ChatGPT users became more indolent, wrote the most dull essays, and showed the least amount of neurological engagement. Many were merely copying and pasting by the end. Despite having no affiliation with MIT and no funding for research, David had come to a similar conclusion.

    LocationDREAM Charter School, East Harlem, New York City
    Grade LevelThird Grade (ages 8–9)
    ActivityStudents drafting their own AI use guidelines and guardrails
    DateApril 2026
    Lesson DesignerLego Education
    Lead TeacherKale Blackshear
    School Co-CEOEve Colavito
    Key Student Quote“AI can turn our brains into mush” — David Ortiz, age 8
    Core Concept Identified by Students“Cognitive offloading” — letting AI do the thinking instead of using your own brain
    NYC DOE StatusStill collecting feedback; finalized AI policy not yet released as of April 2026
    School AI Stats78% of DREAM staff already using AI tools; goal of 60% student adoption as learning aid
    Parent Opposition“Parents for AI Caution” group calling for complete halt to AI in classrooms
    The Third-Grade Experiment: What Happened When Children Were Asked to Govern Their Own AI Rules
    The Third-Grade Experiment: What Happened When Children Were Asked to Govern Their Own AI Rules

    Teacher Kale Blackshear led the DREAM lesson, which was created by Lego Education. Blackshear put it this way: “The school’s job is to get ahead of it rather than pretend otherwise because AI is going to be part of these children’s lives.” Seventy-eight percent of the school’s teaching staff had already used AI tools for planning and preparation by the time the lesson started. The administration was not attempting to outlaw a technology that students were already using at home, on their phones, or on the school-provided devices that are now available in the majority of American classrooms starting in kindergarten. Teaching kids to use the tool without becoming reliant on it was a more difficult objective.

    Karter Nieves, another student, brought up the issue of cheating with a clarity that suggests he had witnessed it firsthand. “A lot of people use AI to cheat on their essays or tests,” he stated. “Sometimes people use it, and they copy and paste.” He was eight years old. He had already identified the pattern, recognized it as a problem, and given it a name. Working on a different school project, a different group of third-graders created their own AI system called “Growth Spurt” and imposed stringent controls on it, preventing it from playing games or providing direct answers. It turns out that kids can write guardrails that are at least as intelligent as the adults who have been discussing the same issues for two years.

    The timing is not coincidental. In late March, the Education Department of New York City published preliminary AI guidelines, which are a traffic-light framework with green for authorized uses, red for prohibited ones, and yellow for situations requiring teacher judgment. Finalized regulations had not yet been released by the chancellor. Parents for AI Caution, a parent organization, advocated for a total halt to the use of AI in classrooms. Education researchers were split; some cited encouraging preliminary findings from specific AI tutoring applications, while others, such as Wayne Holmes, a professor at University College London, pointed out that there is currently no large-scale independent evidence regarding the efficacy or safety of these tools. Both the debate and the uncertainty are real.

    A class of third-graders carrying markers and poster paper entered that gap without waiting. That’s encouraging and a little unsettling at the same time. It is encouraging because it implies that children can think critically about technology in ways that adults sometimes fail to acknowledge when they are given the freedom and space to do so. The fact that eight-year-olds are drafting AI policy in a Harlem charter school while the city’s education department is still gathering input is unsettling because it shows how quickly institutions advance in relation to the rate at which technology enters classrooms.

    In a statement, co-CEO of DREAM Eve Colavito put it bluntly: “Waiting for research risks leaving students to figure this out on their own, without guidance.” That might be precisely correct. It’s also possible that the kids setting cautious boundaries around a technology they hardly understand are more exposed and wiser than any of the adults in the room fully realize, and that moving faster than the evidence supports introduces its own risks.

    At the very least, the East Harlem experiment provided a different kind of beginning, one in which the policy’s most impacted individuals participated in its creation. Education policy typically doesn’t operate like that. Perhaps it’s worthwhile to ask why.


    Disclaimer

    Nothing published on Creative Learning Guild — including news articles, legal news, lawsuit summaries, settlement guides, legal analysis, financial commentary, expert opinion, educational content, or any other material — constitutes legal advice, financial advice, investment advice, or professional counsel of any kind. All content on this website is provided strictly for informational, educational, and news reporting purposes only. Consult your legal or financial advisor before taking any step.

    The Third-Grade Experiment
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Errica Jensen
    • Website

    Errica Jensen is the Senior Editor at Creative Learning Guild, where she leads editorial coverage of legal news, landmark lawsuits, class action settlements, and consumer rights developments and News across the United Kingdom, United States and beyond. With a career spanning over a decade at the intersection of legal journalism, lawsuits, settlements and educational publishing, Errica brings both rigorous research discipline, in-depth knowledge, experience and an accessible editorial voice to subjects that most readers find interesting and helpful.

    Related Posts

    South Korea’s Students Score Highest in the World. Their Mental Health Tells a Different Story

    April 23, 2026

    Inside the Harvard Spinout That Is Disrupting Private Credit and Making Institutional Investors Nervous

    April 23, 2026

    Inside the College That Replaced Half Its Adjuncts With AI — and Called It Progress

    April 23, 2026
    Leave A Reply Cancel Reply

    You must be logged in to post a comment.

    Technology

    Avis’s Data Breach Settlement Is Open for Claims. Here’s What the Hack Actually Exposed

    By Janine HellerApril 23, 20260

    The notice appeared in the mail, nestled between utility bills and grocery flyers, exactly like…

    South Korea’s Students Score Highest in the World. Their Mental Health Tells a Different Story

    April 23, 2026

    Maryland Reaches Mega ‘Settlement in Principle’ With Ship Owner Over Key Bridge Collapse

    April 23, 2026

    Google Updates Gemini Suicide Safeguards as Wave of Wrongful Death AI Lawsuits Mounts

    April 23, 2026

    Designing the Future of Africa: Rice360’s High-Stakes Educational Engineering Competition

    April 23, 2026

    The AI Fluency Index: Anthropic’s New Report Exposes a Massive Global Knowledge Gap

    April 23, 2026

    Oxford Researchers Found That AI Is Making Students Worse at Critical Thinking. Here’s the Evidence

    April 23, 2026

    Shielding Big Oil: Why Republicans Are Rushing to Protect Corporations from Climate Litigation

    April 23, 2026

    The Third-Grade Experiment: What Happened When Children Were Asked to Govern Their Own AI Rules

    April 23, 2026

    Inside the Harvard Spinout That Is Disrupting Private Credit and Making Institutional Investors Nervous

    April 23, 2026
    Facebook X (Twitter) Instagram Pinterest
    • Home
    • Privacy Policy
    • About
    • Contact Us
    • Terms Of Service
    © 2026 ThemeSphere. Designed by ThemeSphere.

    Type above and press Enter to search. Press Esc to cancel.