Interactive mode — your LLM agent drives a live browser through our grid. Autonomous mode — submit a job and walk away.
Scripted flows run at native JS speed; the LLM only wakes up when something deviates.
Two Modes
Interactive mode keeps the LLM on your side. Autonomous mode puts the LLM inside the container, so jobs run unattended.
Your LLM coding agent — Claude Code, Cursor, Copilot — sends JavaScript commands via REST or MCP. The Dashboard proxies them to a Worker container that drives a real browser. No LLM runs on the Dashboard itself.
LLM coding agent
Client-side LLM
Dashboard
Proxy · No LLM
Worker container
host.docker.internal
Session s_8c3b1d
curlA worker-side LLM drives the observe-decide-act loop inside the karate-agent container. Flows run first at native speed — the LLM only takes over for unknown steps. Poll status with GET /api/jobs/$id; download the full report.zip when it completes.
Client → Dashboard
Submit Job
POST /api/jobs
LLM Worker
act
observe · decide
Browser
maxIterations
Dashboard → Client
report.zip
GET /api/jobs/$id/download
maxIterationsDevelopment Workflow
Six steps to move a new app from zero coverage to zero-token regression.
Explore
discover locators
Create
write .js flow
Test
native JS speed
Compose
chain flows
Autonomy
submit, walk away
Report
review · iterate
Explore
Discover locators via interactive mode.
Create
LLM writes .js flow files.
Test
Flow.run() → ok: true
Compose
Chain flows into orchestrators.
Autonomy
Submit as a job, LLM handles unknowns.
Report
Review deviations, fix, repeat.
Architecture
No microservices, no databases, no message queues.
Runtime
Java 21
Container
Docker 24+
Artifacts
1 jar · 1 image
Per Session
~2 GB RAM
API
REST + MCP
Scalability
Same jar, same image — runs on a laptop, a shared server, or dedicated CI infrastructure.
On your laptop. Test against localhost. Flows in a local directory.
Mac Mini or EC2. Dashboard URL shared with the team. Flows backed by git.
Dedicated infra. Jobs submitted by pipelines. Reports as artifacts. Kubernetes for scale.
Real-World Proof
12-step Personal Auto submission — from login through quote creation. Watch the cost collapse as flows take over from the LLM.
Iterations
390
Cost per run
$0.50$0
Run time
4 min15 s
Each step was developed as an independent flow via Interactive mode, then composed into a single orchestrator. The orchestrator runs all 12 flows sequentially at native JS speed. The LLM is invoked only if a step deviates from the expected path.
Guidewire PolicyCenter is a complex enterprise SPA where standard locator strategies fail. Karate Agent's cursor-pointer discovery handles the non-standard <div onclick> targets that Selenium and Playwright struggle with.
Progressive Adoption
Each stage reduces LLM dependency. Most teams land at Stage 3 or 4, where the happy path is deterministic and the LLM only covers edge cases.
Stage 1
~$0.50 · 4 min
LLM explores app from scratch, discovers locators, navigates with no guidance.
Stage 2
~$0.40 · 3 min
Markdown task file provides context. LLM still drives, but with fewer wrong turns.
Stage 3
~$0.02 · 2 min
Scripted flows handle known steps. LLM only on failure.
Stage 4
$0 · ~15 s
Pure deterministic JS. No LLM, no tokens, native speed.
Let us walk you through Interactive mode, flows, and Autonomous mode on your application.