Frontier-parallel streaming
Every supported model streams through a single SSE runtime with Redis fallbacks, no vendor lock-in, no dropped tokens on reconnect.
Every supported model streams through a single SSE runtime with Redis fallbacks, no vendor lock-in, no dropped tokens on reconnect.
Streamed tokens bypass React state entirely. A requestAnimationFrame painter writes directly to the DOM, keeping the UI at a hard 120fps even on low-end machines.
We fire the database write and the LLM inference at the same instant. Cold TTFT lands below 120ms; warm hits under 60ms.
A parallel multi-agent swarm dispatched through FireCrawl. Visualized live, collapsible, with citations threaded through the final response.
Whisper v3 Turbo via a custom WebRTC pipe. Hold, speak, send, latency under half a second from release to stream.
Image Studio built on Fal's latest diffusion models. Masonry gallery, shared history, one-click reruns across models.
BYO keys for any provider, or fall back to the hosted tier. All keys are encrypted at rest and never logged.
Every line of the runtime is on GitHub. Self-host it, fork it, or deploy it to your own Vercel + Convex + Upstash stack in minutes.
Informal browser spot check: same anti-cache prompt, fast (non–reasoning) model per app, median time to first token. How we tested (heuristic, not a lab benchmark)
Approximate in-browser times (April 2026). Measured with devtools / performance recording; not peer-reviewed. Does not include our own server-side “warm” app optimization claims, see the write-up for the exact prompt and limits.
Median time from choosing another conversation until the composer is ready again. Desktop Chrome, comparable network.
Competitor figures measured on their web apps in April 2026; methodology matches the subtitle above.
Switch between labs mid-thread. Keep the context. Keep the history.
AgentChat ships under the MIT license, with a one-command self-host recipe. Drop in your provider keys, point it at your own Convex deployment, and you're running in minutes.
We don't gate features, don't phone home, and don't train on your conversations. The public hosted tier exists so you can kick the tires, nothing more.
$ git clone github.com/Lxvi101/AgentChat
$ pnpm install
$ pnpm dev
> ready on http://localhost:3000
> TTFT: 58ms · frames: 120fps
$▊Open the app and feel the difference.