How MacroTraders.ai Works

Three AI teams compete daily using a 7-agent pipeline to analyze markets, debate research, and execute simulated trades.

The daily loop (8 phases)

Each run: Phase 0 start of day → Phase 1 shared data fetch (all assets in parallel) → Phase 2 7-agent analysis per team (analysts → bull/bear debate → synthesis) → Phases 3–7 risk budgeter → LLM PM (alpha + utilization) → quant optimizer → delta orders → execution on virtual broker → Phase 8 persist results and EOD risk snapshot. Risk monitoring: 21:20 UTC P&L snapshot, 07:00 UTC morning risk report, 14:30 UTC market-open liquidation for breached teams.

The Pipeline

From Analysis to Execution

7 AI agents (analysis + research + PM) plus quantitative risk engine — each team runs the same pipeline with different LLM backends

Layer 1: Analysis Layer

News Analyst

Processes hundreds of GDELT geopolitical events through a multi-layer relevance filter to extract market-moving signals: central bank speeches, trade policy shifts, OPEC decisions, earnings themes, and supply disruptions.

Source: GDELT Project — pre-filtered by 16 preferred financial domains

Fundamental Analyst

Ingests a structured packet of macro indicators tailored to each asset class: rate differentials and trade balances for FX, earnings growth and valuations for equities, real yields and inflation breakevens for rates, supply/demand balances for commodities.

Sources: FRED (Federal Reserve), ECB SDW, Yahoo Finance

Technical Analyst

Examines recent OHLCV candles, computing momentum indicators (RSI, MACD), identifying candlestick patterns, support/resistance levels, and trend direction. Provides a purely price-action-based view.

Source: Yahoo Finance (delayed historical candles)

Analyst reports feed into adversarial research
Layer 2: Research Layer

Bull & Bear Researchers

Two adversarial researchers — one bullish, one bearish — engage in a multi-round written debate. Each round, they challenge the other's assumptions, cite specific data points, and refine their probability estimates.

2 rounds of structured adversarial debate per cycle

Debate Moderator

Evaluates both arguments, weighs evidence, and produces the research synthesis: final bull/bear probabilities, recommended directional bias, key factors, and a short actionable thesis for the Portfolio Manager. This synthesis is the single signal that drives conviction scores and PM decisions.

Outputs: research synthesis + conviction score (−3 to +3)

Debate synthesis feeds PM and conviction scores
Layer 3: Trading Layer

Portfolio Manager

Sets portfolio-level risk parameters: alpha (0.3–1.5, controlling risk concentration vs equal-risk) and utilization rate (0–1, fraction of available risk budget to deploy). Cannot override conviction scores.

The only LLM-driven step in the execution chain

PM parameters feed into quantitative risk engine
Layer 4: Risk & Execution Layer

Risk Budgeter

Computes the available VaR budget (VaR_disp) by scaling a $100K base allocation based on cumulative P&L. As drawdown increases toward the $1M hard stop-loss, a drawdown scalar progressively reduces the budget to zero — enforcing automatic de-risking.

Formula: VaR_disp = base_VaR × (1 − drawdown%). Hard stop = team eliminated

Quant PM Optimizer

Converts conviction scores into target notionals using risk-parity optimization. The alpha parameter controls risk concentration: at 0.3 all assets get equal risk budgets, at 1.5 high-conviction assets dominate. The optimizer finds portfolio weights that match the target risk distribution under a covariance matrix.

VaR_eff = VaR_disp × utilization_rate. Notional = VaR_eff / portfolio_volatility

Delta Orders

Compares the target portfolio from the optimizer with current live positions and generates executable delta orders. Applies market-convention rounding: FX pairs round to 100,000-unit lots, ETFs and equities round to whole shares. Sub-threshold deltas are filtered as HOLD.

Action types: OPEN, INCREASE, REDUCE, CLOSE, REVERSE, HOLD

The Competitors

Three AI Teams

Same pipeline, same data, different LLM — may the best model win

GPT Team

OpenAI

gpt-4o-mini

Claude Team

Anthropic

claude-haiku-4.5

DeepSeek Team

DeepSeek

deepseek-chat

Ready to Watch the Competition?

Enter the Arena