G
GALEA
v0.1 · BRIEF ● LIVE
◆ INVESTIGATION AND OPTIMIZATION FOR AGENT WORKFLOWS

Every agent
workflow needs
an investigator.

Galea plugs into the agent workflows teams already run, follows every event, and turns raw traces into company-aware answers: what happened, what mattered, what was risky, and what should improve next.

MERCURY / PROTON
LANGGRAPH
OPENAI / CLAUDE
CUSTOM RUNTIMES
galea · live trace · session_8a3f
streaming

Agent workflows are becoming
business workflows.

FORCING FUNCTION 01

Failures are no longer toy failures

Agents now refund customers, redline contracts, summarize medical visits, and change product data. When they fail, teams need an answer, not a log dump.

FORCING FUNCTION 02

Every team picked a different runtime

Mercury, LangGraph, OpenAI, Claude, CrewAI, custom queues. The winning observability layer cannot require a rewrite.

FORCING FUNCTION 03

Optimization is customer-specific

Harvey cares about correctness. Decagon cares about refund risk. Cursor cares about unsafe edits. Generic dashboards miss the point.

Teams have traces.
They do not have answers.

Agent runtimes can show events. Existing observability tools can show spans. But product teams still have to manually inspect a workflow and decide whether it was correct, allowed, efficient, safe, and worth changing.

FAILURE 01

Logs without judgment

The trace says a tool ran. It does not say whether that tool should have run for this customer, user, matter, ticket, or policy.

FAILURE 02

Generic priorities

Latency, cost, correctness, compliance, context size, and risky edits are not equally important. Every company weights them differently.

FAILURE 03

Anomalies hide in normal runs

A workflow can finish successfully while using 2x normal tokens, citing unsupported facts, or editing data it should not touch.

FAILURE 04

Incidents do not become improvements

Teams debug one run, then move on. They rarely convert the failure into a reusable eval, baseline, alert, or workflow fix.

Galea sits above
whatever runs your agents.

Keep Mercury, LangGraph, OpenAI, Claude, CrewAI, Temporal, or custom code. Galea listens to the workflow events, builds the timeline, applies company context, and produces the investigation. From the team's point of view, Galea just works.

Product Workflow
support · legal · clinical · coding · ops
Galea Event Layer ◆ JUST WORKS
captures events · traces · tool calls · outputs
Existing Orchestration
mercury/proton · langgraph · openai · custom
Tools · Memory · Models
postgres · pgvector · openai · anthropic · mcp
GALEA CLOUD
trace ingest
company context
investigator agent
anomaly detection
eval recommendations
audit export
Galea does not need to own orchestration to create value. It turns events from any runtime into investigations, alerts, summaries, baselines, and recommended fixes.

Same workflow data.
Different investigations.

Galea is useful because it does not treat every workflow the same. It learns what each company cares about and investigates against that priority model.

Not just traces.
An agent that knows what matters.

Galea's investigator/validator/summarizer walks each workflow with the user. It reads the trace, the company profile, the product surface, prior incidents, and customer priorities, then flags what that team actually cares about: correctness, token usage, context bloat, latency, tool risk, data edits, or anomalous behavior.

The workflow finished.
That does not mean it was good.

Galea looks past success/failure status. It compares each run to the company's baseline, risk model, and product priorities, then explains the part a human should care about.

!
ANOMALY

2.1x normal usage

The run completed, but context grew across three retries and doubled token spend against the customer's normal baseline.

?
CORRECTNESS

Unsupported output

The final answer cited data that was never retrieved. For Harvey, that matters more than latency or cost.

×
TOOL RISK

Edited protected data

The workflow touched a field that normally requires review. Galea flags the run and recommends a guardrail.

2.1x
USAGE SPIKE
3
RISKY TOOL CALLS
1
MISSING SOURCE
4
SUGGESTED EVALS

From incident
to improvement.

The point is not another dashboard. Galea turns each investigated workflow into a reusable improvement: an alert, eval, baseline, policy suggestion, or product fix.

STEP 01

Investigate

Galea explains the workflow in company context
What happened, why it mattered, which event caused the risk, and whether this is anomalous.
↳ human-readable summary
STEP 02

Optimize

Galea recommends the next durable fix
Add an eval, tighten retrieval, alert on a baseline shift, or add a review requirement for risky edits.
↳ workflow-specific recommendation
STEP 03

Monitor

Galea keeps watching the same failure mode
The next time a similar workflow runs, Galea knows whether it improved, regressed, or drifted.
↳ continuous agent QA

A monitoring category,
built for agent workflows.

Galea is not another orchestrator. It is the investigation and optimization layer that works regardless of orchestration layer. Same buyer gravity as Datadog and Sentry, but the unit of analysis is an agent workflow.

COMP 01

Datadog

metrics & traces for cloud servers
MARKET CAP
$40B+
usage-metered, daily-active, retention compounds
↳ monitoring shape
COMP 02

Sentry

error tracking for production apps
MARKET CAP
$3B+
free tier → paid → enterprise
↳ developer adoption motion
COMP 03

LaunchDarkly

feature flags for safe rollouts
MARKET CAP
$3B+
CI gates, deploy safety
↳ safe-change model

Agent workflows need investigation, not just orchestration.

The orchestration layer will vary by team. The need to understand, audit, and improve agent behavior will not. Galea becomes the neutral layer that watches every run, explains what mattered, and turns incidents into better workflows.

◆ THE ONE-LINER
Galea is the investigation layer for agent workflows.