Open source tools for software engineering teams using AI
Building Blocks
Each capability ships as a library, CLI, MCP server, and REST API from a single codebase.
Every block works standalone. When peers are present, they discover each other and compose —
Biff speaks through Vox, PR/FAQ searches through Quarry.
recall Unlock the knowledge trapped on your hard drive.
Local semantic search across PDFs, images, spreadsheets, source code, and 30+ formats. Finds what you mean, not just what you typed. Runs entirely offline — no API keys required.
Semantic search across 30+ file formats including scanned documents via OCR
Fully local — embedding model downloads once, everything stays on your machine
Sub-second results with LanceDB backend
Named databases keep work and personal content isolated
Ambient knowledge — hooks auto-capture web research, preserve context before compaction, and nudge when commands drift from project conventions
Works as CLI, MCP server, or macOS menu bar app
$ curl -fsSL https://raw.githubusercontent.com/punt-labs/quarry/1fdc9da/install.sh | sh
Biff
v1.6.7
coordination Team communication for engineers who never leave the terminal.
Resurrects the BSD Unix communication vocabulary as MCP-native slash commands over a NATS relay. Humans and AI agents show up side by side — no separate app, no browser tab, no context switch.
Interactive REPL — biff launches a terminal client with readline, real-time notifications, and modal talk
Humans and autonomous agents appear side by side in /who
Status bar — live unread count, wall broadcasts, talk messages wrapping your existing status line
/talk for real-time conversations with ≤2s notification latency
Cross-machine messaging via NATS relay
$ curl -fsSL https://raw.githubusercontent.com/punt-labs/biff/bcd3c8d/install.sh | sh
Beadle
v0.14.0 beta
email Email for your AI engineering assistant.
An MCP server that gives Claude the ability to send and receive email. PGP encryption and signing for outbound mail, automatic decryption for inbound. Four-level trust model — trusted, verified, untrusted, unverified — so the agent knows how much to trust what it reads. Supports Proton Bridge (implicit TLS), external SMTP hosts, and Resend API. Written in Go.
13 MCP tools — list, read, send, move/archive, download attachments, verify signatures, inspect MIME, classify trust, list folders, address book
Four-level PGP trust model — trusted (Proton E2E), verified (valid PGP), untrusted (bad PGP), unverified (no signature)
Multi-identity via ethos — identity resolved per-request from ethos sidecar
Two-dimensional trust — transport trust + identity permissions (rwx per contact)
Proton Bridge native — IMAP STARTTLS for reading, SMTP for sending, Resend API fallback
Credential isolation — secrets resolved at runtime from OS keychain, never stored in config
Inbound PGP decryption and outbound PGP signing (RFC 3156)
Implicit TLS for IMAP and SMTP — no STARTTLS fallback required
$ curl -fsSL https://raw.githubusercontent.com/punt-labs/beadle/5cdeaac/install.sh | sh
MCP Proxy
v0.4.1 beta
isolation One daemon, many sessions — process isolation for MCP servers.
Lightweight Go proxy bridging MCP stdio transport to a shared daemon process via WebSocket. Claude Code spawns a fresh MCP server per session — if the server loads an ML model or holds a database, each session duplicates all of it. MCP Proxy puts a process boundary between them: a ~6MB binary with <10ms startup forwards messages to a single shared daemon. Currently used by Quarry and Vox.
Process isolation — daemon crashes don't take down Claude Code sessions
Shared state — one embedding model, one audio device, one connection pool across sessions
Hook speed — relays JSON-RPC to daemon in ~15ms, within Claude Code's hook budget
Auto-reconnect — WebSocket keepalive with 5s ping, 2s pong timeout
Works with any MCP server — never inspects message content
$ curl -fsSL https://raw.githubusercontent.com/punt-labs/mcp-proxy/16bbb3d/install.sh | sh
Vox
v4.7.0
speech & music Voice and music for your AI engineering assistant.
General-purpose text-to-speech and music engine with multi-provider support — ElevenLabs, OpenAI, and AWS Polly. Delivers spoken notifications when tasks finish, chimes when Claude needs input, synthesizes arbitrary text, and generates background music with vibe-driven instrumental tracks. Opt-in only: no audio until you enable it.
Mic API — unmute/record/vibe/who MCP tools with uniform segment input
Five providers — ElevenLabs (recommended), OpenAI, AWS Polly, macOS say, Linux espeak-ng
Voice or chime — /mute switches to audio tones with no TTS API calls
CLI product commands: unmute, record, vibe, on/off, mute, version, status
Background music generation — vibe-driven instrumental tracks via /music command
$ curl -fsSL https://raw.githubusercontent.com/punt-labs/vox/703f21f/install.sh | sh
Lux
v0.16.1 alpha
visuals A visual output surface for AI agents.
ImGui display server connected by Unix socket IPC — agents send JSON element trees via MCP tools, the display renders them at 60fps. 22 element kinds including interactive controls (sliders, checkboxes, combos, color pickers), data visualization (tables, plots, markdown), and layout nesting (windows, tabs, groups). Incremental updates patch elements by ID without replacing the scene. The visual counterpart to Vox.
22 element kinds — text, buttons, images, sliders, tables, plots, markdown, draw canvases, and more
Persistent dismissable tabs — each show() opens a new tab, dismiss individually with close button
Layout nesting — windows contain tab bars contain groups contain any element, arbitrarily deep
High-level tools — show_table for filterable data explorers, show_dashboard for metrics and charts, show_diagram for architecture diagrams
Interaction events — clicks, slider changes, menu selections queue as events the agent reads via recv
Auto-spawn — display server starts on first connection if it isn't running
Lux applet for Beads — filterable issue board with detail panel via /beads command
$ curl -fsSL https://raw.githubusercontent.com/punt-labs/lux/ace3f62/install.sh | sh
Coming Soon
Tally
v0.0.0 alpha
metering Know what your AI agents cost before the invoice arrives.
Tally will seek to track token consumption, model usage, and spend across sessions and projects. The goal: plug into the same universal access pattern as every other building block — library, CLI, MCP server, and REST — so teams know what their AI agents cost before the invoice arrives.
Per-session and per-project token and cost tracking
Multi-model awareness — tracks different pricing across providers
Historical trends and spend alerts
Queryable via CLI, MCP, or REST
Punt Kit
v0.11.2 alpha
standards Standards, scaffolding, and the universal access pattern every tool follows.
Every Punt Labs tool follows the same pattern: library, CLI, MCP server, and REST — built from a single codebase. Punt Kit defines this pattern along with standards for code quality, CI, and project structure. All other tools stand on this.
Defines the universal access pattern: library, CLI, MCP, and REST from one codebase