Skip to main content

OSS voice-to-agent developer terminal

Speak a goal. Route it to memory and execution. Stay in flow.

Capture voice, classify intent, enrich with semantic memory, and launch bounded coding tasks — all from a single desktop terminal built for AI-native development.

OSS desktop app, optimized for solo developers using voice + local AI in daily coding sessions.

agent-terminal://sessioncomposer tab
# 1. clone and start
git clone https://github.com/joshl26/agentterminal
npm install && npm run tauri dev

# 2. open Composer tab, type your goal
# Brain recall enriches context automatically

# 3. pre-send guard validates the request
# then routes to Agent Harness workers

# 4. check Runs tab for live status
5
UI tabs: Composer, Runs, History, Transcripts, Context
125+
integration tests protecting the core session flow
Tauri
native desktop shell on Windows, macOS, and Linux
OSS
standalone delivery before hosted expansion

What it does

Voice, memory, and execution in one terminal — not three tools.

Agent Terminal is the interface layer between your intent and your AI stack. It captures voice, enriches with Brain memory, runs the pre-send guard, and delegates bounded tasks to Harness workers — all from one desktop application you actually keep open.

N

Voice transcript ingestion

Capture speech via Agent Wispr, classify intent, and route transcripts directly into the terminal session without breaking flow.

M

Semantic memory integration

Recall relevant project context from Agent Brain before each session — facts, decisions, and prior work surface automatically.

A

Harness task launch

Compose a goal in the terminal, run the pre-send guard, and delegate bounded coding tasks directly to Agent Harness workers.

C

Persistent session history

Every session and turn is persisted to a local SQLite database. Browse and replay past sessions from the History tab.

D

Native Tauri desktop shell

Runs as a first-class desktop application on Windows (WSL2), macOS, and Linux — not a browser tab.

L

Pre-send guard

A pure, side-effect-free guard evaluates every outgoing message before dispatch — blocking risky or malformed requests before they leave.

How it works

Start with a goal. Add memory and execution only where it helps.

1

Speak or type a goal

Use Agent Wispr to capture voice transcripts, or type directly in the Composer tab. The terminal classifies intent and prepares context.

2

Enrich with memory

Agent Brain recalls relevant project facts, prior decisions, and session history. The enriched context is attached before any task is launched.

3

Run the pre-send guard

Every outgoing message passes through a pure guard function that validates scope, detects risky patterns, and blocks malformed requests.

4

Launch to Harness, review in Runs

Bounded coding tasks are dispatched to Agent Harness workers. Live run status appears in the Runs tab with full attempt history.

Tauri desktop shell
Voice ingestion via Agent Wispr
Semantic memory via Agent Brain
Pre-send guard (pure, no side effects)
Harness task launcher
SQLite session history

First-run shape

Start with `npm run tauri dev`, open the Composer tab, type a goal. Brain recall and Harness delegation are available immediately — add Wispr voice input only when you want hands-free operation.

Positioning

Not another chat UI. A terminal built for the full AI-native loop.

Native desktop shell (not a browser tab)

Agent TerminalYesCursor IDENoWindsurfNoClaude DesktopNo

Voice transcript ingestion

Agent TerminalYesCursor IDENoWindsurfNoClaude DesktopNo

Semantic memory across sessions

Agent TerminalYesCursor IDENoWindsurfNoClaude DesktopNo

Bounded task delegation with verification

Agent TerminalYesCursor IDENoWindsurfNoClaude DesktopNo

Pre-send guard (pure, auditable)

Agent TerminalYesCursor IDENoWindsurfNoClaude DesktopNo

Persistent session history (local SQLite)

Agent TerminalYesCursor IDENoWindsurfNoClaude DesktopNo

Current product path

Standalone OSS first, team operation later.

The roadmap prioritizes reliable daily usage for solo developers before any shared-session or multi-user story.

OSS Desktop

$0

Fully working today. Voice ingestion, semantic memory, Harness task launch, and 5-tab desktop UI on Tauri.

  • YesTauri native desktop shell (Windows/macOS/Linux)
  • YesComposer → pre-send guard → Harness launch flow
  • YesSession history persisted to local SQLite
  • YesAgent Brain memory panel in every session
  • YesVoice transcript ingestion via Agent Wispr toggle
Read Quickstart

Power User

Current focus

Wispr → Terminal toggle and deeper Brain context integration for daily voice-driven coding sessions.

  • YesWispr Settings toggle to enable terminal forwarding
  • YesRicher Brain recall at session start
  • YesImproved intent classification for voice inputs
  • YesTranscript deduplication and smart routing
  • YesKeyboard-first shortcuts for power users
View Roadmap

Team-Ready Later

Planned

Shared session history, multi-user Brain workspaces, and remote Harness controller pairing — after the solo path is locked in.

  • YesShared session and transcript history
  • YesMulti-user Brain workspace support
  • YesRemote controller pairing with Harness
  • YesExport and backup flows
  • YesSolo mode always remains first-class
Read Docs

FAQ

Practical answers before platform mythology.

Is Agent Terminal open source?

Yes. The current product direction is standalone OSS first, with a Tauri desktop build as the canonical supported path.

What problem does it solve?

It bridges voice input, semantic memory, and AI task delegation into one desktop terminal — so you can speak a goal and have it routed to the right tool without leaving your workflow.

Do I need Agent Brain and Agent Harness?

No, but the terminal is designed to work best with both. Brain provides memory enrichment; Harness handles bounded task execution. Each is independently optional.

What is the best first run?

Clone the repo, run `npm install && npm run tauri dev`, open the Composer tab, and type a goal. Check the Runs tab to see the Harness response.

Does it require Wispr for voice input?

No. Agent Wispr is an optional integration. You can type directly in the Composer tab and get the full memory-enriched, guard-protected, Harness-delegated flow.

Where should I start if I want to evaluate it quickly?

Use the install guide and quickstart docs. They walk through the Tauri build, first session, and Harness pairing in under 15 minutes.

Get your first voice-to-agent session running in under 15 minutes.

Clone the repo, build the Tauri shell, type a goal in the Composer tab, and see the pre-send guard and Harness delegation in action.