8 New Deep Research Proposals
8 New Deep Research Proposals
Category: research
Date: 15 February 2026, 17:36 UTC
Original File: PROPOSED_DEEP_RESEARCH_PROMPTS.md
Based on Claude reports 1 & 2 + research portfolio
Proposed Deep Research Prompts
Based on Findings from Reports 1 & 2 + Your Research Portfolio
PROMPT C: The Detection Tool Equity Crisis
Research Question: How do AI detection tools perform across different English language proficiency levels, and what are the implications for international students in UK higher education?
Background: Liang et al. (2023) found systematic bias against non-native English writers. Your portfolio includes detection tool evaluations. This extends that work with focus on equity.
Specific Investigation:
- Replicate Liang et al. methodology with current detection tools (Turnitin, Originality.ai, GPTZero, Copyleaks)
- Test with student writing samples from:
- Native English speakers (UK, US, Australia)
- Non-native speakers (Chinese, Arabic, Spanish, Hindi first-language)
- Different proficiency levels (IELTS 6.0, 7.0, 8.0)
- Measure false positive rates by language background
- Analyze linguistic features triggering misclassification
- Document consequences for students (anxiety, appeals, academic consequences)
Data Sources:
- International student writing corpora (public datasets)
- Detection tool APIs
- University case studies (anonymized)
- Student interviews (if ethically approved)
Output: Policy brief on equitable assessment + academic paper
PROMPT D: The Essay Mill Legislation Enforcement Gap
Research Question: Why have UK, Australian, and Irish essay mill laws produced virtually no prosecutions, and what enforcement mechanisms would be effective?
Background: Report 1 noted legislation enacted but almost no prosecutions. Your 2018-2020 work mapped the contract cheating industry.
Specific Investigation:
- Document all legislation:
- UK Skills and Post-16 Education Act 2022
- Australian TEQSA legislation
- Irish provisions
- Any others enacted 2022-2026
- Interview/key informant:
- TEQSA (Australia) enforcement data
- UK Department for Education
- University legal/compliance officers
- Why no prosecutions?
- Analyze successful prosecutions in other jurisdictions (if any)
- Model effective enforcement:
- Criminal vs civil penalties
- Extraterritorial jurisdiction
- Payment processor blocking
- ISP blocking (as TEQSA does)
- Compare to successful enforcement in similar areas (essay mills vs piracy, etc.)
Output: Policy recommendations + comparative law analysis
PROMPT E: The Performance-Learning Paradox Replication Study
Research Question: Is the “performance-learning paradox” (higher homework scores, lower exam scores with AI use) replicated across UK universities, and what assessment designs prevent it?
Background: Report 2 found this pattern at Stanford and CMU. Your teaching module puts you in position to investigate this directly.
Specific Investigation:
- Obtain anonymized grade data from:
- Imperial College modules (your own teaching data)
- Partner UK universities (Russell Group)
- Pre-2023 vs 2023-2026 comparison
- Correlate with:
- Assessment type (exam vs coursework)
- AI policy permissiveness
- Student self-reported AI use
- Identify assessment designs that:
- Resist AI assistance
- Maintain validity with AI available
- Protect learning outcomes
- Cost-benefit analysis of redesign
Output: Best practice guide for assessment design + journal article
PROMPT F: Humanizer Tools Efficacy Arms Race
Research Question: How effective are “humanizer” tools (Undetectable AI, StealthWriter, HIX Bypass) at evading current detection systems, and what countermeasures exist?
Background: Report 1 mentioned these tools exist but limited systematic testing. Your technical background enables rigorous evaluation.
Specific Investigation:
- Systematic testing protocol:
- Generate text with GPT-4, Claude, Gemini
- Process through 5+ humanizer tools
- Test against 5+ detection tools
- Measure detection rates before/after humanizing
- Document humanizer techniques:
- Synonym replacement
- Sentence restructuring
- “Perplexity/burstiness” adjustment
- Adversarial patterns
- Test detection countermeasures:
- Can detectors be trained to catch humanized text?
- What linguistic features persist?
- Track tool evolution over 6 months
Ethical Note: Document for defensive purposes only (improving detection, policy)
Output: Technical report + detection tool improvement recommendations
PROMPT G: International Student Conference Meta-Analysis
Research Question: What do 3 years of student-led academic integrity research reveal about student perspectives on cheating, AI, and integrity that faculty research misses?
Background: Your student conferences (2023-2025) contain underutilized data. Student voice often more credible to peers.
Specific Investigation:
- Transcribe and code all 3 conference videos
- Thematic analysis:
- What topics do students choose?
- How do they frame integrity? (positive/negative)
- What solutions do they propose?
- How has this changed 2023→2025?
- Compare to faculty research:
- Different questions asked?
- Different methodologies?
- Different conclusions?
- Identify student research suitable for:
- Journal publication (with student co-authors)
- Policy recommendations
- Further development
- Create student research showcase
Output: Meta-analysis paper + student co-authored publications
PROMPT H: CS Curriculum AI Integration Comparison (Student Outcomes)
Research Question: Which of the 26 university approaches to GenAI in CS education produces the best student learning outcomes, and what can others learn from them?
Background: Report 2 documented 26 different approaches. Now test which work.
Specific Investigation:
- Partner with 5-10 universities from the 26
- Collect outcome data:
- Exam performance (AI-resistant assessment)
- Project quality
- Graduate employment
- Student satisfaction
- Academic integrity violations
- Correlate with approach:
- Prohibition vs integration
- Infrastructure investment
- Assessment redesign
- Training provided
- Cost-effectiveness analysis
- Identify transferable best practices
Output: Evidence-based recommendations for CS educators
PROMPT I: Multilingual Academic Integrity Discourse
Research Question: How do discussions of academic integrity, contract cheating, and AI differ across languages and cultures on social media?
Background: Your portfolio includes Twitter analysis. Extend to multilingual + your multilingual framework.
Specific Investigation:
- Collect social media data:
- English (UK, US, Australia)
- Spanish (Spain, Latin America)
- Chinese (Weibo, Zhihu if accessible)
- Arabic (Twitter MENA)
- Other languages as possible
- Code themes:
- Attitudes toward cheating
- Views on AI use
- Trust in institutions
- Suggested solutions
- Compare:
- Cultural differences in integrity concepts
- Different AI adoption patterns
- Varying policy preferences
- Translate key insights
Output: Cross-cultural analysis + policy implications for international education
PROMPT J: Assessment Redesign Efficacy Database
Research Question: Which specific assessment redesign strategies successfully maintain integrity and learning outcomes in the GenAI era?
Background: Report 2 noted shift to assessment redesign but limited evidence of what works.
Specific Investigation:
- Systematic review:
- Identify all papers on assessment redesign 2023-2026
- Extract redesign strategies
- Categorize by discipline, level, assessment type
- Efficacy coding:
- Which maintain integrity?
- Which protect learning?
- Which are feasible at scale?
- Cost to implement?
- Create database:
- Assessment type
- Redesign strategy
- Evidence of efficacy
- Implementation guidance
- Gap analysis:
- What hasn’t been tried?
- What needs evaluation?
Output: Searchable database + best practice guide
RECOMMENDED PRIORITY ORDER
Immediate (Start This Week):
- C - Detection Equity (builds on Liang et al., high policy relevance)
- G - Student Conference Analysis (data already exists, unique contribution)
Short-term (Next 2-4 Weeks):
- E - Performance-Learning Paradox (uses your teaching data)
- F - Humanizer Tools (technical, timely)
- D - Legislation Enforcement Gap (policy impact)
Medium-term (1-2 Months):
- H - CS Curriculum Outcomes (requires partnerships)
- I - Multilingual Discourse (scale up gradually)
- J - Assessment Redesign Database (systematic review)
SELECTION CRITERIA
Each prompt selected based on:
- ✅ Builds on your prior research
- ✅ Addresses current critical issue
- ✅ Feasible with available data/tools
- ✅ High impact potential
- ✅ Publication opportunity
- ✅ Conference presentation value
- ✅ Student engagement potential
Which 2-3 should we start immediately?
Original file: PROPOSED_DEEP_RESEARCH_PROMPTS.md