Pilot live: ACP for AI commerce.Explore ACP
Forecasting agent

Forecasts that learn. Every prediction compounds.

Felix predicts revenue eight weeks ahead, retrained nightly on your accounts. Confidence narrows over time as he learns your business.

Why this exists

Forecasting that breaks every Monday. And never gets sharper.

Models go stale instantly.

An analyst builds a forecast on Monday. By Wednesday, the inputs have shifted. By Friday, nobody trusts the numbers. Repeat weekly.

Cycles take weeks, not minutes.

Pulling data from five platforms, cleaning it, running regressions, sanity-checking. 15 to 20 hours per forecast cycle that could be 15 seconds.

Misses do not improve next time.

Last quarter's forecast was off by 18%. Did the model learn? No. Same model, same blind spots, same miss. Just with newer data.

Trends arrive late, not early.

Seasonal shifts, competitor moves, creative fatigue. A spreadsheet cannot detect emerging trends three weeks before they hit revenue.

What Felix ships

Five capabilities. One forecasting workflow.

Each capability maps to a specific job: predict the line, learn from the miss, surface the trend, alert on the anomaly, and stay aware of the rest of the workforce.

8-week revenue forecasts

Channel-level predictions with confidence intervals that narrow as Felix learns your business.

  • Per-channel and per-campaign breakdowns
  • Confidence ranges, not single numbers
  • Updated nightly on your accounts

Example. $420K-$460K next month at 87% confidence

Compound learning

Every prediction gets scored against reality. Felix dissects misses and never makes the same one twice.

  • Backtest history kept per account
  • Miss attribution: which feature drove the gap
  • Same mistake never compounds across cycles

Example. 78% accuracy on day one, 91% by week eight

Early trend detection

Spots emerging patterns two to three weeks before they show up in your dashboard.

  • Seasonal shifts, demand inflections
  • Creative fatigue and channel decay
  • Competitor pressure picked up via Parker

Example. Hook fatigue flagged 18 days before performance dropped

Anomaly alerts

When reality diverges from forecast, Felix surfaces it within 24 hours, not at month-end close.

  • Spend, conversion, and CAC anomalies
  • Confidence-band breach alerts to Slack
  • Routes to Parker for attribution sanity check

Example. Spend spike at 6:47 AM, resolved by noon

Workforce-aware

Reads Maya's brand context, Sam's scenarios, Dana's reconciled data. No silo, no manual reconciliation.

  • Sam scenarios priced in seconds, not hours
  • Maya's CAC ceiling factored into ranges
  • Dana feeds clean data; no garbage in

Example. Sam tested a 30% Meta shift; Felix priced it in 4 seconds

A typical week

What Felix does while the team sleeps.

One workweek, hour by hour. Specific numbers vary by account.

Mon · 4:32 AM
Trained

Runs nightly retrain on the last 12 months of account data. Eight models in ensemble, one prediction per channel.

Mon · 5:14 AM
Forecast

Forecasts revenue 8 weeks out. Channel-level ranges, confidence narrows on Meta and Google, widens on TikTok.

Mon · 7:30 AM
Flagged

Detects creative fatigue forming on Hook_v4. Routes to Olivia for variant generation before performance drops.

Mon · 9:18 AM
Anomaly

Anomaly: spend up 34% overnight. Cross-checks with Parker. 23% real lift, 11% platform inflation.

Mon · 11:02 AM
Scored

Sam asks for a +30% Meta shift scenario. Felix prices the outcome in 4 seconds with confidence bands.

Tue · 5:10 AM
Reconciled

Re-scores last week's predictions against actuals. Off by 4% on Wednesday; logs the miss for next cycle.

Tue · 8:00 AM
Delivered

Delivers the weekly recap to Slack. Three forecasts logged with confidence bands, two anomalies surfaced.

Compound accuracy

78% on day one. 98% over time.

Forecast accuracy is not a static number. Felix's predictions compound against your business, and the gap closes every cycle.

Week 1

78%

Cold start on your historical data. Baseline ensemble across ARIMA, Prophet, Greykite. Already sharper than spreadsheet forecasts.

Week 4

85%

First prediction cycles get scored against actuals. Felix identifies which features drove the gap and adjusts weights.

Week 8

91%

Compound effect kicks in. Cycles reinforce each other. Confidence bands narrow on stable channels.

Week 24+

98%

Mature on your account. Predictions read as decisions, not guesses. Maya holds the why, Felix holds the what.

ARIMA + Prophet + Greykite ensemble, nightly retrain

What's underneath

Real infrastructure. Not a wrapper.

Felix runs on a purpose-built forecasting stack: ensemble models, nightly retraining, scored backtests, and shared memory across the workforce.

Reads from

Meta, Google, TikTok, and Shopify via OAuth. Dana feeds reconciled, deduped data so Felix never trains on noise.

Meta AdsGoogle AdsTikTok AdsShopify

OAuth-based, no warehouse copy

Reports to

Slack for daily anomaly alerts, Sheets for weekly forecast tables, Notion for the long form. Format adapted per surface.

SlackGoogle SheetsNotion

Anomaly alerts within 5 minutes of breach

Collaborates with

Maya supplies brand memory and CAC ceilings. Parker validates spend lift. Sam runs scenarios. Dana keeps inputs clean.

Shared memory, not message-passing

Built on

ARIMA for seasonality, Prophet for trend changepoints, Greykite for the long horizon. Eight models in ensemble, one confidence-scored output.

Nightly retraining, scored backtests every cycle

Ready when you are

See Felix forecast your account, or talk to us about your stack.

The pilot connects your accounts to all seven agents. Felix's first forecast finishes in under five minutes.

Looking for something specific? Parker validates lift, Dana keeps inputs clean or Sam tests scenarios.