AI in Education VIP Research Exchange
A collaborative knowledge base for NYU's AI in Education VIP team.
VIP Subteams & Research Areas
Understanding, predicting, and guiding the impact of generative AI in education through four complementary research pathways:
Evaluating Tools 0
Understanding Users 0
Assessing Impacts 0
Innovating EdTech 0
Background Resources
Learning Science 0
Research Methods & Resources 0
Understanding Users Attitudes & Behaviors
These resources help us to answer questions like: How, and how much, are students and teachers using AI tools? What values, principles, and assumptions shape their engagement with AI tools?
General 0
General perspectives on how people perceive and interact with AI systems.
Administrators 0
Administrative perspectives on AI policy, implementation, and institutional change.
Students 0
Understanding how students perceive and use AI tools in their learning processes.
Faculty 0
Faculty attitudes toward AI integration and their teaching practice adaptations.
General Perspectives on AI
General perspectives on how people perceive and interact with AI systems.
Faculty Perspectives on AI
Faculty attitudes toward AI integration and their teaching practice adaptations.
One of the most interesting findings was that the more teachers recognize ethical issues in AI, such as fairness, the more they become knowledgeable about the integration of AI-based tools in education. This caused them to expand the TPACK model into Intelligent-TPACK, highlighting the importance of ethics in a teacher's role in AI integration. I thought this was fascinating because I think teachers want what's best for their students, and learning about these technologies to learn how to decide if they can be used fairly in their classrooms seems like such a great angle for raising interest. They also found that offering professional development that taught teachers how to use AI would help them come up with ways that AI could enhance their own established pedagogical practices. These tools offer a lot of room for innovation, and I think giving teachers the knowledge they need to be able to innovate on their own methods offers a much less scary relationship to AI than being told to use it in a specific way. At one point, the paper says that Intelligent Technologies should be regarded as more of a partner than a tool, and I think a lot of the TPACK principles can help teachers to learn to see it that way. The study concludes by hoping for future work that tests whether the Intelligent-TPACK framework can be used as a predictor of the quality of teachers' use of AI tools. There is a lot of focus on testing on the student side, so it was very interesting to see the development of a test for the teacher side.
Student Perspectives on AI
Understanding how students perceive and use AI tools in their learning processes.
Administrator Perspectives on AI
Administrative perspectives on AI policy, implementation, and institutional change.
I started liking the concept of the learning module (7) on prompt engineering - this is where I thought it would be leaning more toward what I was hoping for which is more about effectively using AI in learning and especially not just asking for a generation of body of work. This topic I felt this proposal failed at. Though, the rest of the modules, to me, felt to extensively and adequately cover policy, ethics, basic computer science, understanding models and engineering, and other more meta socio-perceptions and projections of AI.
CONCLUSION: Overall great opening to the conversation and foundation to being thinking of how to structure learning about AI - OPPORTUNITY to skew and expand upon it even more toward education and creating - - aka how can you teach about the learning and creating process with AI 😁?
More resources on administrative perspectives coming soon!
Student Attitudes & Behaviors
Understanding how students perceive and use AI tools in their learning processes.
Evaluating AI Tools
These resources help us to answer questions about the capabilities of generative AI tools, which tools perform better on certain tasks, and how well AI-detection works.
Coding and Math 0
Evaluating AI tools' capabilities in programming, code generation, and debugging tasks.
Creativity 0
Assessing AI's ability to generate creative content and support creative thinking processes.
Detection Tools 0
Evaluation of AI detection technologies and their effectiveness in educational contexts.
Ethics / Safety 0
Examining ethical considerations and safety concerns in AI tool deployment.
Equity / Social Justice 0
Analyzing how AI tools impact educational equity and social justice issues.
Error / Hallucination 0
Understanding AI errors, hallucinations, and accuracy limitations.
Reasoning 0
Testing AI's logical reasoning, problem-solving, and critical thinking capabilities.
Usability 0
Assessing user experience, interface design, and ease of use in AI tools.
Translation 0
Evaluating AI translation capabilities and multilingual support quality.
Tutoring / Instruction 0
Evaluating AI tutoring systems, instructional capabilities, and educational feedback quality.
Detection Tools
Evaluation of AI detection technologies and their effectiveness in educational contexts.
This paper recommends shifting school work to in-person centered assessments, presentations, and having on-going check-ins for writing. This could limit the use of AI and lead to students writing on their own. Another approach to this issue, the paper suggests having transparent disclosures on the proper uses of AI. Having these set rules will help properly integrate AI into learning. As the paper mentions, AI mirrors the invention of the calculator, as like AI, it was at first looked down upon and considered cheating and banned in schools. However, a gradual shift in perception occurred where calculators were later accepted in schools and they began regulating its use and now see it as a tool to make mathematics more efficient. I think once we are past this period of AI slander and begin to see its potentials, creating clear and transparent guidelines for its use will bring down the uses for cheating and bring down our reliance on unreliable AI Detection tools.
The researchers also measured the performance of three AI detection tools (Google's Bard, ZeroGPT, and GPTZero) on this task. The research found that human-written abstracts had significantly higher (less predictable) perplexity scores than their AI-generated counterparts (AUC=0.7794).
The primary utility of this research is its direct comparison of commercial detection tools, which revealed a massive disparity in performance. GPTZero (an AI detector that also uses perplexity) was found to be highly accurate (95%), while ZeroGPT was mediocre (69%) and Google's Bard performed worse than chance (36% accuracy). This finding is important for educators, administrators, and journal editors, as it demonstrates that while AI detection is possible, the reliability of a given tool cannot be assumed (this also supports the findings of some of the other articles under the Evaluating Tools section).
Furthermore, this study provides a good empirical test case for the other articles. It applies the core concept of perplexity that Chiusano explains, but its findings are immediately complicated by other research (like Lindberg & Nobel's and Liang's) showing this exact metric is unreliable, biased against non-native speakers, and insufficient without other features like "burstiness." This research also directly benchmarks the same tools (GPTZero and ZeroGPT) that Pratama, Weichert, and Elkhatat evaluate. While Elek finds GPTZero is highly accurate (95%), these other papers provide the crucial counterpoint that it is also biased, easily fooled by prompt engineering, and inconsistent against newer models. This highlights the real-world consequences of imperfect detection, such as the "imposter bias" and the genre-specific false positives.
The research's strength is in demonstrating a significant vulnerability in existing detection systems. The study shows that prompts generated by SICO enabled GPT-3.5 to successfully bypass six different detectors, significantly outperforming baseline paraphrasing methods. A comprehensive human evaluation further confirmed that the SICO-generated text maintained human-level readability and task completion rates.
This work is valuable because it highlights that evasion can be achieved through simple prompt-guiding rather than external tools, and it provides a new, effective benchmark (SICO) for evaluating the robustness of future detectors.
It also connects well with several past papers I've found that looks into prompt engineering and using that to bypass AI-detection tools (through targeting perplexity and burstiness).
Equity / Social Justice
Analyzing how AI tools impact educational equity and social justice issues.
Reasoning
Testing AI's logical reasoning, problem-solving, and critical thinking capabilities.
Ethics / Safety
Examining ethical considerations and safety concerns in AI tool deployment.
Submodel for educators:
- Equity concerns
- Digital literacy
- Enhance vigilance of bias
- Pedagogy and assessment concerns
- Encourage higher-order thinking
- Incentivize the learning process
- Strike a balance
- Optimizing potential of technology
- Establish a general policy
- Set specific boundaries
- Update policy regularly
Submodel for students:
- Preparation for using GAI
- Improve awareness of limitations
- Carry-out critical evaluation
- Applying GAI
- Seek out additional resources
- Incorporate creativity and originality
- Disclosure
- Acknowledge and site sources properly
I found that the specific breakdown of responsibility for both educators and students to be well organized and thought out - I felt that the formatting gave a clear understanding of intention and a real attempt to find solutions and answers to questions which remain fairly open ended. I don't know if the literary introduction is as indicative of what the paper really seeks to argue - however I appreciate the reference and the greater thought framework it represents.
Usability
Assessing user experience, interface design, and ease of use in AI tools.
The article critiques how increasing machine agency threatens human autonomy through privacy erosion, algorithmic opacity, automation bias (overtrusting machines), and algorithm aversion (rejecting algorithms even when optimal). Users often feel helpless against incomprehensible algorithms, experiencing "algorithmic anxiety" (Airbnb hosts) or unawareness of curation (Facebook feeds). To address this, the author advocates for human-AI collaboration: seeking user assent before automated decisions, increasing algorithmic transparency, enabling users to train and direct algorithms ("ground-truthing"), and pursuing mutual augmentation where humans and AI extend each other's capabilities. Trust built through deeper engagement (action route) proves more robust than perception alone (cue route), ultimately calling for co-creation that respects human agency while leveraging machine capabilities.
The article critiques the AI and computer science communities for developing systems in isolation without adequate input from HCI professionals and behavioral scientists. It criticizes how machine behavior is primarily studied by those without formal behavioral science training, how "explainable AI" is often built for AI professionals rather than end users, how ML engineers are positioned as the "human at the center" instead of actual users, and how ethical guidelines lack practical implementation methods. The authors advocate for a Human-Centered AI (HCAI) approach where humans remain the ultimate decision makers, calling for HCI professionals to: leverage interdisciplinary expertise to translate psychological theories into computational models, develop human-in-the-loop systems that combine human and machine strengths, create meaningful human control mechanisms for autonomous systems, ensure AI explainability leads to true user comprehension, and apply iterative design methods with end users throughout AI development rather than as an afterthought. The paper emphasizes that effective AI systems require deep collaboration between HCI and AI professionals from the earliest stages of development.
Coding and Math
Evaluating AI tools' capabilities in programming, code generation, and debugging tasks.
Content in Development
VIP Evaluating Tools subteam is currently curating resources on AI coding capabilities. We welcome your contributions and findings.
Creativity
Assessing AI's ability to generate creative content and support creative thinking processes.
Content in Development
VIP Evaluating Tools subteam is currently curating resources on AI creativity and creative content generation. We welcome your contributions and findings.
Error / Hallucination
Understanding AI errors, hallucinations, and accuracy limitations.
Translation
Evaluating AI translation capabilities and multilingual support quality.
Content in Development
VIP Evaluating Tools subteam is currently curating resources on AI translation capabilities. We welcome your contributions and findings.
Tutoring / Instruction
Evaluating AI tutoring systems, instructional capabilities, and educational feedback quality.
I think Zhang makes valid points about not relying on AI to completely change the way we write. The convenience is definitely tempting especially when you're stuck and don't know where to start. However, I think that true skill when it comes to writing comes from struggling to get something down, constant revisions, and hearing what other real humans have to say about it. I think this is where our creativity can really come out. I think that in the end, while ChatGPT and AI are useful tools that can definitely be integrated in some ways and forms, it is best for us humans to continue to learn how to develop our writing ourselves in the traditional way that we know of.
These responses, or the lack of a response, are saved and analyzed in order to recommend follow-up tutoring or enrichment tasks based on inferred strengths and weaknesses. All of these factors go into creating a personalized system for each individual student, which the student can go home and interact with like a one-on-one tutor. Feedback and interaction are based on student history and knowledge gaps, and students can engage in two-way conversations through SMS with their tutor. Teachers also have dashboards, which allow them to guide and initiate tutoring interventions.
DARTS uses secure cloud-based databases with encrypted data transmission and storage, and all data is handled in compliance with FERPA. Using cloud-native design, the system is scalable to classrooms of any size, and through SMS, the system can be used even without an internet connection.
These papers are a three-part series, and the third part will look at actual classroom integration. I think this is a fascinating set of articles, and I am very interested in technology like this. I think there are many things that need to be extensively researched and tested before something like this could be rolled out, but I feel very optimistic about the possibilities that this gives to allow for Mastery Learning in a school system that doesn't accommodate those who can't master the content at the set pace. The article, like my previous two articles, stresses the importance of teacher training for integrating technology like this and introducing students to the technology at the beginning of the school year.
Assessing Impacts
These resources help us to answer questions like: What are the effects of AI tools on student learning outcomes? How do these tools impact academic integrity, assessment validity, and student wellbeing? What are the equity implications of AI adoption in education?
Learning Outcomes 0
Measuring how AI tools affect student learning, comprehension, and skill development.
Academic Integrity 0
Assessment of AI's effect on academic honesty and institutional integrity measures.
Assessment Validity 0
How AI tools impact the reliability and validity of educational assessments and grading practices.
Student Experience + Wellbeing 0
Impact of AI on student stress, motivation, confidence, and overall educational experience.
Equity + Access 0
How AI tools affect educational equity and access across different student populations.
Learning Outcomes
Measuring how AI tools affect student learning, comprehension, and skill development.
The authors create an example ITS called the Socratic Playground for Learning (SPL), which implements this system. Through a user interface, students are able to interact with these models by defining, through a menu or a text box, what they are learning about or practicing. The student then engages in back-and-forth Socratic-style conversation. The questions presented to them dynamically scale in difficulty and complexity depending on their answers and responses.
To test this system, the authors involved 30 first-year students from the Faculty of Education at Central China Normal University, all majoring in Early Childhood Education. In order to measure the system's effectiveness quantitatively, all of these students were studying beginner English, and they were provided with a pre- and post-test before and after a week of using SPL for 30-40 minutes each day. The students were also provided with a survey asking questions grouped into effectiveness, engagement, adaptivity, satisfaction, and recommendation activity. By the end of the week, there was a significant improvement on the post-test. Effectiveness, engagement, and adaptivity were rated very high on the qualitative side, with recommendation accuracy just slightly below, and satisfaction received good but lower scores. The study showed the efficacy of the system, though there is no control group so it is hard to fully call it a causal relationship.
As the authors point out, it's interesting that the satisfaction scores were lower, and that highlights the importance of determining how to create systems that students actually want to use. It's unsurprising that they provided results, but figuring out how to design these systems in ways that can not just become tools for efficiency, but student-desired tools for better experiences in their education is paramount. My last review was about another ITS system that used text as an interface, and while one of the major issues of SPL is that the use of ChatGPT like this makes the system not applicable in lower resource environments, it's likely that a text-based system would score even lower on satisfaction. I will be interested in continuing to explore this idea of how to make these systems integrate into classrooms in ways that give students useful personalized attention, but don't detract from or change for the worse pedagogical practices that exist now.
It finds that 95.6% of students use AI in academic activities through virtual assistants (88.2%) and educational platforms (42.4%). Most of the participants use these tools weekly or daily, making AI an integral part of the educational process for tasks like homework, projects, and knowledge enhancement. While 80% of respondents agree that AI enhances their educational experience and 82.4% believe it improves academic performance, some express uncertainty or skepticism about its ultimate effects, including concerns about reducing critical thinking and accuracy issues. When asked for suggestions on improvement, most students suggest proper integration of AI in education, focusing on guided use that doesn't limit critical thinking. Overall, the majority of participants highlight both the positive influence on collaboration and learning efficiency, as well as challenges such as the need for reliable information and ethical use.
This paper gives valuable insight into the real-time experiences and perceptions of students who have direct exposure to AI-integrated learning environments, rather than relying solely on expert or educator perspectives. By using mixed-methods analysis across both closed and open-ended survey questions, it provides a nuanced account of how students engage with artificial intelligence for academic activities, covering not just usage statistics but also the emotional, ethical, and cognitive challenges and benefits they encounter. The study emphasizes student-driven recommendations for balanced and effective AI integration, including calls for enhanced training, reliability, and critical engagement. This makes it a key resource for educators and policymakers aiming to understand and responsibly advance AI adoption in education, learning from the people who would be most affected by these policies.
What I thought was important to note in this paper was how it cited the need to include emotional and behavioral cues in AI, adapting to the learning environment based on the individual student's needs; this particular emphasis is because merely adjusting the difficulty of tasks based on student performance is insufficient to become a personalized tutor. Since we as humans need more than just efficient learning tools, but also crave more connection, it is important to recognize that GenAI has a long way to go before even threatening to replace teachers in the classroom.
Overall, the paper highlights the need for equitable AI use across diverse educational contexts: rural and urban schools, as well as among students with disabilities. Ensuring transparency and accountability in AI decision-making is crucial to building trust and promoting inclusive adoption of these technologies in education.
However, the paper stresses that education shouldn't just be about memorizing facts or passing standardized tests. Instead, schools need to cultivate "Learning Mastery"—helping students understand how they learn best, think critically, set goals, and reflect on their progress. Alongside this, "Knowledge Mastery" means achieving a deep, conceptual understanding of subjects, not merely recalling information, so students can transfer their knowledge to real-world challenges.
The article emphasizes the importance of ethical AI deployment, warning against issues like bias, privacy risks, and inequality. It calls for collaboration between educators, policymakers, and technologists to ensure AI benefits all students. Teachers need professional development focused on both using AI and nurturing the unique human qualities—creativity, empathy, independent thinking—that future societies and workplaces will need. In conclusion, schools must combine the best aspects of AI and human intelligence to prepare students for lifelong learning and success in an AI-driven world.
Innovating EdTech
This section catalogs existing AI-powered educational tools to help innovators understand the current landscape and identify market gaps. Unlike other sections focused on research evaluation, these pages describe tools and their capabilities to inform product development and innovation strategy.
AI Tutoring & Learning Support 0
Tools that act as instructors, mentors, or study aids including tutorbots, homework helpers, and adaptive learning systems.
Writing & Content Generation Tools 0
Tools that help students create or refine text and media, including writing assistants and translation tools.
Detection Tools 0
Tools that monitor or verify the authenticity of work, including AI-writing detectors and plagiarism checkers.
Organizational & Productivity Tools 0
Tools that structure learning and manage large projects, such as note-taking and research assistants.
Feedback & Assessment Tools 0
AI that evaluates or comments on student work, including automated grading and feedback generators.
Accessibility & Inclusion Tools 0
AI that expands educational access through speech-to-text, captioning, and assistive technologies.
Classroom & Learning Analytics 0
AI systems that track and interpret educational activity, including analytics dashboards and early warning systems.
Scheduling 0
Tools that facilitate appointment and meeting scheduling for educational contexts.
Prototype Development
Creating and iterating on educational technology prototypes that leverage AI.
AI Tutoring & Learning Support
Tools that act as instructors, mentors, or study aids including tutorbots, homework helpers, and adaptive learning systems.
Teachers struggle to find time to create personalized learning plans for each student; students receive generic instruction that doesn't match their learning preferences.
A generative learning planner for higher education that creates AI-taught courses based on student preferences. Teachers upload material and teaching structure/vision to create learning plans; students receive personalized plans tailored to their needs.
Teachers and students in higher education
Web application
Freemium
B2B2C (institutions → students)
Coursera, Khan Academy, traditional LMS platforms
Specifically designed for higher education with dual-sided personalization—teachers define structure/vision while students receive individually tailored plans. Founded by NYU ITP post-graduate.
Feedback & Assessment Tools
AI that evaluates or comments on student work, including automated grading and feedback generators.
Teachers spend excessive time on tedious administrative tasks (grading, feedback, recommendation letters, IEP goals) that could be better spent on instruction.
A web/browser extension with functions including writing inspection, feedback & grading, presentation making, IEP goal generator, and recommendation letter generator.
Teachers and students (K-12 focus)
Browser extension (Chrome), integrates with Google Docs, Slides, Classroom
Free for teachers
B2C / B2B (free individual, school plans)
Magic School AI, Diffit, Eduaide.ai
Browser-based approach that works within existing workflows (Google Suite). Exhaustively presents AI possibilities through extensive function menu—a "quick option" effect that shows teachers AI uses they wouldn't think of themselves.
Organizational & Productivity Tools
Tools that structure learning and manage large projects, such as note-taking and research assistants.
Teachers spend hours creating quality worksheets, activities, and slide-decks instead of focusing on direct instruction.
Generates course assets including worksheets, activities, assessments, and slide-decks.
Teachers (K-12 focus)
Web application
Freemium (limited free tier, paid plans)
B2C / B2B (freemium individual, school plans)
Magic School AI, Canva for Education, Google Slides + AI tools
Focused specifically on asset generation (worksheets, slides, assessments) rather than broader AI assistant functions. Higher education applicability unclear—primarily K-12 oriented.
Teachers are overwhelmed by lesson planning and IEP writing; school administrators need AI tools they can trust with student data.
Lesson planning, assessments, and IEP writing capabilities with 60+ AI tools for educators.
Teachers (K-12 focus)
Web application, Chrome extension
Free tier, paid school/district plans
B2B / B2C (school/district focus, free individual tier)
Brisk AI, Almanack, Eduaide.ai, Diffit
Heavy emphasis on "safeguards," evidence basis, and data security—strong trust-building messaging. Comprehensive suite approach with 60+ specialized tools. Higher education applicability unclear—primarily K-12 oriented.
Writing & Content Generation Tools
Tools that help students create or refine text and media, including writing assistants and translation tools.
Writers struggle with paraphrasing, grammar, and refining their work; students need writing assistance that helps them improve rather than writes for them.
AI writing-specific tool that provides pop-up style recommendations as you write, including paraphrasing, grammar checking, and summarization.
Teachers, students, and general users
Web app, browser extension, MS Word add-in
Freemium (limited free, premium subscription)
B2C (individual subscriptions)
Grammarly, Wordtune, ProWritingAid, Hemingway Editor
Markets itself as a collaborator rather than "writing for you"—emphasizes responsible AI use. Inline recommendations appear as you write. Strong paraphrasing focus. Inclusive but not specifically tuned for formal academic settings.
Detection Tools
Tools that monitor or verify the authenticity of work, including AI-writing detectors and plagiarism checkers.
Content in Development
VIP Innovating EdTech subteam is currently curating resources on detection tools. We welcome your contributions and findings.
Accessibility & Inclusion Tools
AI that expands educational access through speech-to-text, captioning, and assistive technologies.
Content in Development
VIP Innovating EdTech subteam is currently curating resources on accessibility and inclusion tools. We welcome your contributions and findings.
Classroom & Learning Analytics
AI systems that track and interpret educational activity, including analytics dashboards and early warning systems.
School administrators lack visibility into AI usage across their institution; teachers need AI tools while administrators need oversight and analytics on student/teacher performance.
Comprehensive platform with administrative oversight capabilities including rule setting and insight to student/teacher performance. Interface includes quick options like rubric generator, multiple choice quiz builder, and text translator.
Teachers, school leaders, students, and administrators
Web application
Free for teachers, paid district/school plans
B2B (school/district contracts)
Magic School AI, Khanmigo, traditional LMS platforms with AI add-ons
All-in-one platform where students, teachers, and administrators coexist. Has higher education tab. Quick options give teachers AI ideas. However, raises concerns about surveillance and data collection when building infrastructure with administrative oversight of all users.
Scheduling
Tools that facilitate appointment and meeting scheduling for educational contexts.
Content in Development
VIP Innovating EdTech subteam is currently curating resources on scheduling tools. We welcome your contributions and findings.
Faculty Perspectives
Faculty attitudes toward AI integration and their teaching practice adaptations.
Content in Development
VIP Understanding Users subteam is currently curating resources for this research area. We welcome your contributions and findings.
Administrator Views
Administrative perspectives on AI policy, implementation, and institutional change.
Content in Development
VIP Understanding Users subteam is currently curating resources for this research area. We welcome your contributions and findings.
AI Competencies & Literacy
Assessment of AI literacy levels and competency development across user groups.
GenAI Tool Capabilities
Assessment of what current generative AI tools can and cannot do for education.
Tool Limitations & Biases
Understanding the constraints, biases, and failure modes of AI educational tools.
Comparative Tool Analysis
Side-by-side comparisons of different AI tools for specific educational tasks.
Content in Development
VIP Evaluating Tools subteam is currently curating comparative studies. We welcome your tool evaluation research and findings.
Academic Integrity
Assessment of AI's effect on academic honesty and institutional integrity measures.
The research's strength is in demonstrating a significant vulnerability in existing detection systems. The study shows that prompts generated by SICO enabled GPT-3.5 to successfully bypass six different detectors, significantly outperforming baseline paraphrasing methods. A comprehensive human evaluation further confirmed that the SICO-generated text maintained human-level readability and task completion rates.
This work is valuable because it highlights that evasion can be achieved through simple prompt-guiding rather than external tools, and it provides a new, effective benchmark (SICO) for evaluating the robustness of future detectors.
It also connects well with several past papers I've found that looks into prompt engineering and using that to bypass AI-detection tools (through targeting perplexity and burstiness).
This paper recommends shifting school work to in-person centered assessments, presentations, and having on-going check-ins for writing. This could limit the use of AI and lead to students writing on their own. Another approach to this issue, the paper suggests having transparent disclosures on the proper uses of AI. Having these set rules will help properly integrate AI into learning. As the paper mentions, AI mirrors the invention of the calculator, as like AI, it was at first looked down upon and considered cheating and banned in schools. However, a gradual shift in perception occurred where calculators were later accepted in schools and they began regulating its use and now see it as a tool to make mathematics more efficient. I think once we are past this period of AI slander and begin to see its potentials, creating clear and transparent guidelines for its use will bring down the uses for cheating and bring down our reliance on unreliable AI Detection tools.
Assessment Validity
How AI tools impact the reliability and validity of educational assessments and grading practices.
Student Experience + Wellbeing
Impact of AI on student stress, motivation, confidence, and overall educational experience.
Equity & Access Issues
How AI tools affect educational equity and access across different student populations.
User Testing & Feedback
Methods for testing educational prototypes with real users and gathering feedback.
Integration Strategies
Approaches for integrating new AI technologies into existing educational systems.
Content in Development
VIP Innovating EdTech subteam is currently curating integration methodologies. We welcome your implementation research and case studies.
Scalability Assessment
Evaluating the potential for scaling educational technology innovations.
Research Methods & Resources
These sources will help you identify a research gap, design appropriate research methodologies, interpret your findings, and find publication in a peer-reviewed journal.
Literature Review Methods 0
Systematic approaches to finding, evaluating, and synthesizing existing research in AI education.
Collaborative Research 0
Best practices for team-based research, project management, and interdisciplinary collaboration.
Experimental Design 0
Methods for designing controlled studies to test AI tool effectiveness and educational interventions.
Survey Design & Analysis 0
Best practices for creating effective surveys and analyzing quantitative data in educational research.
Interview & Focus Group Methods 0
Techniques for conducting qualitative interviews and focus groups with students, faculty, and administrators.
Data Analysis Techniques 0
Statistical and qualitative analysis methods for educational research data and findings interpretation.
Research Ethics & IRB 0
Ethical considerations and IRB approval processes for educational research involving human subjects.
Academic Writing & Publication 0
Guidelines for writing research papers, conference presentations, and academic publications.
Literature Review Methods
Systematic approaches to finding, evaluating, and synthesizing existing research in AI education.
Content in Development
VIP Research Methods subteam is currently curating resources on literature review methodologies. We welcome your contributions and findings.
Collaborative Research
Best practices for team-based research, project management, and interdisciplinary collaboration.
Content in Development
VIP Research Methods subteam is currently curating resources on collaborative research practices. We welcome your contributions and findings.
Experimental Design
Methods for designing controlled studies to test AI tool effectiveness and educational interventions.
Content in Development
VIP Research Methods subteam is currently curating resources on experimental design. We welcome your contributions and findings.
Survey Design & Analysis
Best practices for creating effective surveys and analyzing quantitative data in educational research.
Content in Development
VIP Research Methods subteam is currently curating resources on survey design and analysis. We welcome your contributions and findings.
Interview & Focus Group Methods
Techniques for conducting qualitative interviews and focus groups with students, faculty, and administrators.
Content in Development
VIP Research Methods subteam is currently curating resources on interview and focus group methods. We welcome your contributions and findings.
Data Analysis Techniques
Statistical and qualitative analysis methods for educational research data and findings interpretation.
Content in Development
VIP Research Methods subteam is currently curating resources on data analysis techniques. We welcome your contributions and findings.
Research Ethics & IRB
Ethical considerations and IRB approval processes for educational research involving human subjects.
Content in Development
VIP Research Methods subteam is currently curating resources on research ethics and IRB processes. We welcome your contributions and findings.
Academic Writing & Publication
Guidelines for writing research papers, conference presentations, and academic publications.
Content in Development
VIP Research Methods subteam is currently curating resources on academic writing and publication. We welcome your contributions and findings.
Learning Science
Background readings in the science of teaching and learning.
Writing & Composition 0
Cognitive and developmental theories of writing processes and instruction.
Cognitive Science 0
Research on cognitive load, memory, creativity, and learning processes.
Reading & Literacy 0
Deep reading, textual interpretation, and the ethics of literary engagement.
Educational Philosophy 0
Philosophical foundations of education, purpose, measurement, and teacher agency.
AI/ML Foundations 0
Technical foundations of AI and machine learning relevant to education.
Writing & Composition
Cognitive and developmental theories of writing processes and instruction.
Cognitive Science
Research on cognitive load, memory, creativity, and learning processes.
Reading & Literacy
Deep reading, textual interpretation, and the ethics of literary engagement.
Educational Philosophy
Philosophical foundations of education, purpose, measurement, and teacher agency.
AI/ML Foundations
Technical foundations of AI and machine learning relevant to education.
Ongoing Projects
Current research initiatives by the AI in Education VIP team.
Domain-Specific AI Detection at NYU
We are building a domain-specific AI detector trained on Congressional Research Service reports.
Most existing detectors are "all-purpose" and fail under adversarial conditions. By focusing on a narrow domain (student essays, starting with Congressional Research Service reports), we aim to develop a detector that is more accurate and interpretable than general tools like GPTZero/Turnitin.
And perhaps the most exciting part: we won't just train on better data but we'll also explore architectural improvements to the transformer itself, pushing beyond standard models to capture structural differences between human and AI writing that current detectors miss.
NYU Writing Center Platform
This project continues the development of an online writing platform that bridges first-year and upper-level writing by embedding expert strategies from EXPOS courses into an accessible platform.
Designed as a "reinforcement point," it helps students carry forward effective writing habits beyond first-year writing. Unlike traditional word processors, it both (a) promotes the kind of meta-reflection that motivates radical revision and (b) provides tools for enacting those revisions.
In this way, the program demystifies the habits of expert writers, supports lasting skill development, and reduces the attractiveness of AI shortcuts.
Our Team
Our VIP team is composed of dedicated students and faculty contributing to AI in Education research.