Hi, I'm Sowmith.👋

I build theinfrastructurethat takes AI fromdemo to production.

At Fractal.ai I'm building the harness around production agents — tool registry, evals, the plumbing around the prompt.

Before that, ML infra at Amazon — pipelines, serving, and deployment for inspection models running on millions of packages a day.

The stack is the moat.

Ship the harness, not the prompt.

If it's not measured, it's not real.

Guardrails are a feature.

From demo to production

A demo is one prompt that works. Production is the session, harness, sandbox, eval, and guardrails that make it work again tomorrow when the model changed.

Model drift — an LLM's behavior shifts as providers ship new versions. Prompts, tool calls, and evals built on one model may not survive the next.

Chaos

gpt-5.3-codexClaude Opus 4.7Llama 4Responses APIAssistants API (deprecated)MCP server (half-built)pgvectorprompt_v12.mdscratch.ipynb.env.localagent.pyTODO: eval harnessTODO: retry logicuseless_log.txtLangChain

Orchestration

  • Session

    Append-only event log + memory

  • Harness

    The loop: plan, call, retry, exit

  • Tool Registry(MCP)

    Typed, discoverable, permissioned tools

  • Evals & Guardrails

    CI for model swaps

Production

  1. User
  2. Router
  3. Harness
  4. Response

Agentic infrastructure is the layer between the model and the user — session, harness, tool registry, evals. It's where most of the engineering cost of an AI product now lives.