Building the Future of AIET IN STEM: A Working Group Discussion

Chris Quintana Knut Neumann Yasemin Copur-Gencturk Umesh Ramnarain Mei-Hung Chiu + more participants
Back to Blog

What happens when researchers from six continents gather to discuss the future of AI in STEM education? A rich debate about frameworks, conferences, journals, equity, and how to create something truly different from the academic status quo.

Not Just Another Conference

This working group session moved beyond typical conference planning to ask fundamental questions: What unique value can this community create? How do we bridge theory and practice? And how do we ensure the Global South has a real seat at the table?

A Global Framework for AI in STEM Education

The Big Opportunity

This group has the potential to develop a global framework for AI in STEM education - guidelines, principles, and evaluation criteria that don't exist yet but are urgently needed.

Unlike siloed academic conferences, this group brings together diverse perspectives - science education, math education, AI researchers, practitioners, and international viewpoints. This diversity is exactly what's needed to create something comprehensive.

  • Guidelines for thinking about and using AI in education
  • Watch-outs and risk considerations
  • Evaluation frameworks for AI tools
  • Ethical guidelines that companies currently lack
View original discussion

Participant: "It seems to me that this group has the potential to do something bigger. If it's not developed yet, why couldn't this be a working group to put out that product for a global basis? Here's a framework, here's some rules, here's some guidelines about how we should be thinking about and using AI in education."

Rethinking Academic Publishing

The Speed Problem

AI changes every day. If we wait six months to a year for reviewer comments, the research is already outdated. Traditional journal publishing is too slow for this field.

Why Not Another Journal?

The group debated creating a new journal but raised important concerns: it takes years to establish credibility, there are already many journals, and the traditional model doesn't fit AI's pace of change.

Experienced editors in the group shared hard-won wisdom about journal publishing challenges:

  • Getting into citation indexes takes "forever"
  • Impact factor matters for career evaluation in many countries
  • Reviewer recruitment is already difficult across existing journals
  • Running a journal is massive work for people with full-time jobs
View original discussion

Knut Neumann: "Don't we already have enough journals? There's AI&Ed, Journal of Learning Analytics, JRST, and all these journals. The other thing that typically is underestimated is how long it takes to establish a journal, partly because it takes you forever to get into the social scientific citation index."

The Alternative: Proceedings + Dissemination

Instead of a traditional journal, the group converged on conference proceedings (like ICLS) combined with new forms of dissemination - use cases from teachers, practical examples, video demonstrations.

The vision is something different from "yet another journal":

  • Proceedings that are treated as real publications
  • Use cases and practical examples from teachers
  • Point/counterpoint discussions
  • Video demonstrations alongside written papers
  • Digital publications that truly convey educational contexts
View original discussion

Chris Quintana: "I wouldn't mind thinking about some sort of dissemination vehicle for things - use cases from teachers, practical things that show here's what people are trying to do with AI. Point counterpoint discussions. Something different than yet another journal."

Reimagining the Conference Model

The Core Question

If I'm going to add another conference to my limited travel budget, there has to be a reason to select this over others. What makes this different?

Why STEM Matters (Not Just Science or Math)

NARST is science-focused. NCTM is math-focused. There's no conference that brings STEM disciplines together around AI - and that cross-pollination is where the real learning happens.

Participants shared their experiences at discipline-specific conferences and the value of cross-disciplinary exchange:

  • "In the last two days I learned way more than I learned in all the math conferences I attended"
  • Math educators and science educators face similar AI challenges but rarely talk
  • Bringing disciplines together doesn't mean ignoring discipline-specific aspects
  • Some challenges (like students bypassing learning with ChatGPT) matter across all disciplines
View original discussion

Yasemin Copur-Gencturk: "I really like the idea of we call it in NCTM 'linking research to practice.' It's a three-day conference, the third day always there is a connection to the practice. Teachers, practitioners join, and there are learning opportunities for teachers. I don't think there is a conference for STEM education."

Bringing Different Worlds Together

A conference that only attracts researchers talking to researchers is "yet another conference." The vision that emerged brings together four key stakeholder groups:

Researchers
Practitioners
Designers
AI Companies
The AI Company Challenge

AI companies have no IRB, no ethics boards. Their AI is often unsupervised. They need to hear about the risks - and researchers need to understand the technical possibilities.

The group discussed both the risks and opportunities of engaging AI companies:

  • Companies may use proxy data (like zip codes) that encode socioeconomic bias
  • Many edtech companies "have no idea about research" and sell based on teacher quotes alone
  • But bringing companies in exposes them to important ethical considerations
  • If we're selective, being invited becomes a privilege that signals quality
View original discussion

Janice Gobert: "Companies are not thinking about that. They have no ethics, they have no IRB. Their AI is unsupervised - it could take in zip code to determine edge cases, which in the United States is a proxy for socioeconomic status."

Participant: "Maybe that's exactly why it would be good to bring the companies in - to hear those struggles and see what the risks are."

Being Selective: Theory + Data Required

If companies want to present, they need to answer: What's the pedagogy underlying your system? What's your data? You can't just say "teachers love it."

The consensus was that selectivity creates value - but the criteria should be clear:

  • Theory-driven: What pedagogical framework underlies your approach?
  • Data-driven: What evidence supports your claims?
  • Development phase: Where are you in validation?
  • Not just "everyone is welcome" - meaningful vetting creates prestige
View original discussion

Yasemin Copur-Gencturk: "If we are selective, then it could be privileged for them to be part of this. It signals something. If we create a very good vetting system - who can come, who can present - it creates value. The standing out part: we are data-driven, we are theory-driven."

What's Missing in Current Conferences

The AI&Ed Problem

AI in Education (AI&Ed) conferences are dominated by Carnegie Mellon, have a "coarse-grained idea about learning," and treat all domains as equivalent - they think "learning science is like learning math is like learning French."

Experienced participants shared observations about existing conferences:

  • AI&Ed is heavily CMU-influenced with "unified theories of cognition" assumptions
  • Context and domain-specific learning isn't prioritized
  • Learning sciences conferences have rich theory but aggregate to holistic scores
  • There's no single conference combining rich theory, deep technology, and practical application
View original discussion

Janice Gobert: "I go to AI&Ed pretty regularly. It's extremely dominated by Carnegie Mellon. I wouldn't say it's theoretically impoverished, but it's not theoretically rich. They have a very coarse-grained idea about learning. They think learning science is like learning math is like learning geography is like learning French. Context isn't that important to them. That's where there could be room for bringing together rich theoretical, practical, and computational work."

The Global South Perspective

A Critical Voice

Umesh Ramnarain from South Africa raised a fundamental question: "Africans have always been sidelined in these conversations. What is my role in this? What benefit do we get back?"

Not Passive Listeners, Active Stakeholders

The Global South has unique perspectives to share - working within severe resource constraints has bred innovation. But participation must be meaningful: research collaboration, capacity building, not just attendance.

The discussion revealed both challenges and opportunities:

  • AI tools often come with prices inaccessible to Global South institutions
  • African researchers want to be developers, not just consumers
  • Low-cost implementation strategies are valuable to everyone
  • Research collaboration and capacity building for emerging scholars is critical
  • 70% of online images are white males - training data bias is a real problem
View original discussion

Umesh Ramnarain: "We are not looking for handouts. We want to be part of the process. We want to contribute. The unfortunate reputation is that Africans are excellent consumers but not developers. We're trying to change that narrative. We work with a very young team of researchers - if there are opportunities to be pulled into development teams and mentored by senior developers, that would make an incredible contribution."

Knut Neumann: "The Global South has something to share that the Global North can learn from. Germany could learn from South Africa about establishing equity with AI. And regarding representation - it's utterly important. If we're thinking about this conference being global, it can't just be a Northern Hemisphere conference."

The Bias Problem in AI

Before 2015, facial recognition performance differed by 33% between white males and black females. The biases are now more subtle but still embedded in LLMs trained on imbalanced data.

The discussion surfaced important AI bias considerations:

  • Early examples were obvious: "doctor + man - woman = nurse"
  • Modern LLMs have more subtle biases worked into their responses
  • Algorithms can distinguish boy vs girl answers with high accuracy
  • Mitigation approaches help but don't eliminate bias
  • Global South participation helps identify problems others miss
View original discussion

Jiliang Tang: "Before 2015, the facial recognition performance difference for white men and black females was 33%. It's because black females joined the efforts to change the situation. This problem is partly from training data imbalance, but at the algorithm level, we can solve this. That's why it's very important to join the global efforts - if you're left behind, nobody would even find this problem."

Emerging Consensus

Not a Journal

Conference proceedings with new dissemination formats - video, demos, practical use cases

STEM, Not Silos

Cross-disciplinary exchange between science, math, engineering, and tech educators

Multiple Stakeholders

Researchers + practitioners + designers + companies (with clear vetting)

Truly Global

Global South as active partners, not passive recipients

Quality Standards

Theory-driven and data-driven requirements for all presentations

Framework Development

Creating global guidelines for AI in STEM education

The Vision

"If we're going to go through the trouble of going to another thing, it's going to be different. We've got enough standard conferences." — Chris Quintana