Multi-Agent LLM Experimentation Framework

Overview

Designing scalable multi-agent LLM experimentation frameworks to study causal effects in identity disclosure and stance change, targeting ACL 2026 submission on identity-conditioned persuasion dynamics.

Key Technologies

  • LLMs: 12 frontier models (GPT-4, Claude, Gemini, etc.)
  • Automation: Playwright, Chromium/Electron orchestration
  • Experimentation: Controlled R1-R2 role structures, Coordinator-LLM consistency scoring

Research Contributions

  • Designed experiments across 12 frontier models with 1.9K sessions
  • Built Chromium/Electron orchestration layer unifying browser chatbot sessions and direct API inference
  • Developed controlled R1-R2 role structures for studying identity disclosure effects
  • Implemented coordinator-LLM consistency scoring at scale

Technical Implementation

  • Browser Automation: Playwright automation for parallel prompting
  • Response Capture: DOM-based extraction from web interfaces
  • API Integration: Unified interface for direct model API calls
  • Consistency Scoring: LLM-based evaluation of multi-agent interactions

Research Goals

  • Study causal effects in multi-agent LLM interactions
  • Analyze identity-conditioned persuasion dynamics
  • Understand stance change in controlled conversational settings
  • Target publication: ACL 2026

Advisor

Prof. Mingfeng Lin, Georgia Institute of Technology