Arbor connects to your engineering tools, pulls real signals across the review period, and uses AI to produce a structured evidence brief for every engineer. You make the call.
Every brief is the same shape: a calibrated summary, the underlying metrics, the trends behind them, and the trajectory across periods.

“Reviews get written the weekend before they're due, from whatever the manager remembers.”
The last six weeks dominate a six-month review. Steady contributors get under-credited; recent firefighters get over-credited.
GitHub has every PR, every review, every comment. Jira has the tickets, the cycle times, the delivery patterns. The story is there. Nobody has time to read it.
Managers walk into committees with a few anecdotes and a vibe. Calibration drifts toward whoever speaks loudest, not whoever has the strongest evidence.
Managers can't see the work clearly. Engineers can't see the decision at all. We asked people across consulting, banking, big tech, and startups how they're actually reviewed. Five different processes, the same underlying disease.
“The visible process is one thing. The real process is calibration, and we never see it.”
“Manager's opinion only. Performance calls are just formality.”
“You're assigned a coach from another team who fights your case at review. How good they are matters more than the work.”
“Structured on paper. Manager's opinion in practice.”
“Whoever you ask for feedback matters more than what you actually did.”
Quotes from individual contacts describing their employer's process. Lightly edited for clarity. Not affiliated with or endorsed by any company named.
Arbor is hosted. Connect once during onboarding, then kick off a review run whenever the cycle comes around.
GitHub and Jira at launch, with more on the way. Personal access tokens or app-style integrations, whichever your security team prefers. Read-only access only, tokens encrypted at rest.
Pick a cohort: a team, a sub-team, an org. Pick a date range. Arbor handles the cohort calibration math.
One brief per engineer: deterministic metrics, an AI-powered narrative grounded in citations, cohort-relative context. Read in-app or export the structured data.
Arbor's output is a structured document the manager and calibration committee read together. Every section is grounded in evidence the team can re-open.
PR throughput, code review depth, cycle time, ticket health. Every number is computed from raw events you can audit. No black-box scoring.
Arbor turns hundreds of events into a paragraph that reads patterns, not platitudes. Every claim links back to the PR, comment, or ticket it came from.
Engineers are compared within their team, over the same period, doing comparable work. The cohort is the calibration; nothing gets stacked against an external benchmark.
Arbor produces the structured artifact your calibration committee anchors on. The brief never outputs a rank or recommendation; that judgment stays human.
Arbor is hosted to the same constraints your platform team would impose. Read-only access, encrypted credentials, no model training on your data.