The agent CLI that doesn't need the internet.
Whatever model server your team already runs.
Most teams have a centralized inference server on their own GPUs: vLLM or TGI behind an internal endpoint, sometimes routed through a governance gateway. Meshly CLI points at that endpoint. Workstation Ollama works too if you're solo. Anything that speaks OpenAI-compatible HTTP. The CLI handles the agent loop; you choose the brain and where it lives.
- Shared inference: vLLM, TGI, or any OpenAI-compatible server on your GPU cluster.
- Through your gateway: route via your existing model proxy or governance layer.
- Workstation: Ollama or llama.cpp on a laptop for solo / offline work.
Same muscle memory. Different brain.
Meshly CLI mirrors the Claude Code workflow: same tool surface (Read / Edit / Write / Bash / Glob / Grep), same task-oriented agent patterns, same MCP server extension model. It does not call api.anthropic.com. It does not speak the Anthropic Messages API. It just gives your engineers the muscle memory they already have against the model server you point it at.
- Workflow-compatible, not protocol-compatible. The Anthropic API isn't in the picture.
- Same tool surface as Claude Code. Same prompts, same patterns, same MCP servers.
- Engineers who use Claude Code switch over in a session, not a sprint.
No phone-home. No public-internet calls.
The only outbound call Meshly CLI makes is to the inference endpoint you configure. No telemetry, no auto-update pings, no anonymous usage stats, no 'check for new release' beacons. Bring it in on a USB drive, run it on a closed network, sleep at night.
- Zero public-internet requirement. All traffic stays inside your perimeter.
- No telemetry, no metrics export, no crash reporting to anyone but you.
- Reproducible offline installs: vendored dependencies, signed binaries, deterministic builds.
Works standalone. Better together.
Meshly CLI is a complete coding agent on its own. When you also use Meshly Build Server, the CLI dispatches tasks from the shared queue, streams output back to the dashboard, and pushes architecture scans for the team. Same role as Station, just terminal-native.
- Standalone mode: your inference endpoint, your repo, no Server required.
- Connected mode: same task / agent / audit surface as Station, driven from a terminal.
- One config file. Switch between standalone and connected without reinstalling.
For the teams that can't ship code to cloud AI.
Cloud-AI coding agents work great until your compliance officer reads the data processing addendum. Meshly CLI is the answer for environments where the cloud isn't an option. Bring the agent to the data, not the other way around.
CLI runs the agents where the cloud can't go.
Available by request.
Meshly CLI isn't a public download. Tell us about your environment, what models you run, what restrictions you operate under, and we'll work out whether it fits and get you a build with the right configuration.