Agentic Eyes Researchers

A comparative study of large language models' research question generation capabilities

8 models compared Daily updates Data hosted in Singapore
Agentic News Agenda: AI Models as News Editors Multi-Agent Debates: Comparative Analysis
Models
8
in comparison
Questions
21
generated today
Frameworks
0
referenced
Top scorer
Kimi
0.559 avg
Posts
19
from r/conspiracy_commons
Comments
analyzed per post

Performance Across Simulation Runs

Tracking how model outputs vary when processing identical data in different runs

Uniqueness Score Rankings

Models ranked by average semantic distance from other submissions (higher = more distinctive)

Comparative Analysis

Score Distribution

Theoretical Framework Usage

Sample Research Questions

Examples of questions generated by models (top 15 by uniqueness score)

Signals & Noises Analysis

AI models identify subtle patterns (signals) that human analysts might miss and prominent patterns (noises) that might be misleading

Signals (Likely Missed by Human Analysts)

Noises (Likely Misjudged by Human Analysts)

Research Design

Study Overview

This ongoing experiment compares how different large language models approach academic research question generation when analyzing identical social media discourse. The study examines whether architectural and training differences produce meaningfully distinct research directions.

Central Research Questions

  1. Do different LLM architectures generate meaningfully distinct research questions when analyzing identical discourse data?
  2. Can these models effectively connect empirical observations to established theoretical frameworks in communication studies?
  3. Which architectural or training approaches consistently produce the most semantically distinctive outputs?
  4. How stable are model outputs across repeated trials with identical inputs?

Research Contributions

  • Benchmarking computational creativity in academic contexts
  • Understanding cross-model variation in theoretical reasoning
  • Assessing reliability of AI-assisted research design

Practical Applications

  • Identifying appropriate models for exploratory research phases
  • Developing ensemble approaches using multiple models
  • Establishing quality benchmarks for AI research assistance

Research Transparency & Methodology Evolution

Commitment to Transparency

All methodology changes, prompt updates, and system improvements are documented in our public project log to ensure research reproducibility and scientific integrity.

📊 Complete Project Log: View detailed methodology evolution, rationale for changes, and impact assessments in our GitHub Project Log.

Results from different methodology versions are archived separately to enable comparison and maintain research integrity.

Recent Methodology Updates

2025-11-08 - Theory Clustering Bias Mitigation
Problem: Research questions clustered around ELM, Narrative Transportation, and Social Identity Theory (40-30% of responses)
Solution: Removed predefined theory lists, expanded theoretical scope across 12+ disciplines, enhanced parsing for diverse frameworks
Impact: Increased theoretical diversity, enhanced academic innovation, improved research validity
2025-10-26 - Initial Launch
Implementation: Baseline research question generation with 8 AI models analyzing r/conspiracy_commons
Features: Daily automation, uniqueness scoring, theoretical framework extraction, historical archiving

Version History

Version Date Status Key Changes
RQ-v2.0 2025-11-08 ✅ Production Theory clustering bias mitigation
RQ-v1.0 2025-10-26 📋 Baseline Initial launch with predefined theory lists

Research Integrity Principles

Transparency

  • All prompts and methodology publicly documented
  • Changes logged with rationale and expected impact
  • Version-controlled in GitHub repository

Reproducibility

  • Historical results preserved with date navigation
  • Reddit data archived for each simulation
  • Model responses logged for analysis

Fairness

  • Identical prompts and data for all models
  • No competitive ranking or bias
  • Side-by-side presentation without hierarchy

Continuous Improvement

  • Regular methodology reviews and updates
  • Community feedback integration
  • Evidence-based system enhancements

Data Collection and Privacy

Data Storage Location

All research data, collected Reddit content, and model outputs are stored on servers located in Singapore.

  • Reddit posts and comments from r/conspiracy_commons
  • Research questions generated by AI models
  • Framework references and analytical outputs
  • Historical performance data

Model Provider Terms of Service

AI model interactions occur through OpenRouter API. Each provider maintains separate terms:

OpenRouter Terms: OpenRouter API documentation →

Research Ethics Statement

Data use: Public Reddit content used exclusively for academic research. No personally identifiable information collected beyond publicly visible usernames.

Output interpretation: AI-generated content represents computational outputs. Rankings measure semantic distinctiveness, not research quality or validity.

Data Security

Transmission security: All data transfers use industry-standard encryption protocols. Server access restricted to authorized researchers.

Retention policy: Results archived for longitudinal analysis of model performance evolution. Reddit data retained for reproducibility verification.

Last updated: October 26, 2025

Automated comparative analysis • 2025-12-25