io.github.spuentesp/lain
High-performance knowledge graph MCP server for AI coding agents with semantic search.
Ask AI about io.github.spuentesp/lain
Powered by Claude Β· Grounded in docs
I know everything about io.github.spuentesp/lain. Ask me about installation, configuration, usage, or troubleshooting.
0/500
Reviews
Documentation
LAIN-mcp
LAIN builds a map of how all the code in your project connects β what calls what, what depends on what, which files tend to change together. Then it lets your AI coding assistant ask questions about that map. So instead of the AI just looking at one file and guessing, it can ask "if I change this function, what else breaks?" and get a real answer. It plugs into any AI agent that supports MCP and runs in the background while you work.
TL,DR:
# One-line install (interactive - will ask you to configure)
curl -fsSL https://raw.githubusercontent.com/spuentesp/lain/main/install.sh | bash
# Add to PATH
export PATH="$HOME/.local/lain:$PATH"
# Or manual install with options (non-interactive)
curl -fsSL https://raw.githubusercontent.com/spuentesp/lain/main/install.sh | \
bash /dev/stdin --workspace . --transport both --yes
What is Lain?
Lain is a persistent code-intelligence MCP server. It builds a queryable knowledge graph of your codebase β symbols and their relationships extracted via LSP and tree-sitter, augmented with git co-change history and optional semantic embeddings β and exposes that graph through MCP tools. The value over LSP-only or RAG-based approaches is cross-file structural reasoning: blast radius for proposed changes, transitive dependency traces, anchor identification, co-change correlation, and contextual build failure decoration so agents can reason about callers rather than just the failing line. Written in Rust, persists across sessions, stays fresh during editing via a file watcher that updates a volatile overlay layered on top of the static graph.
Installation
Quick Install (recommended - interactive)
curl -fsSL https://raw.githubusercontent.com/spuentesp/lain/main/install.sh | bash
The installer will ask you to configure:
- Workspace path
- MCP transport mode (stdio, http, or both)
- HTTP port (if using http/both)
- Target agent (auto-detects Claude Code, Cursor, Windsurf, Cline)
- Whether to download the ONNX model for semantic search
After you confirm your settings, it will:
- Download and install LAIN to
~/.local/lain - Optionally download the ONNX model (~120MB)
- Run
lain initwith your configuration - Add LAIN to your agent's settings
Non-interactive install (with options):
# Install with specific workspace and download ONNX model for semantic search
curl -fsSL https://raw.githubusercontent.com/spuentesp/lain/main/install.sh | \
bash /dev/stdin --workspace . --transport both --download-model --yes
# Install for specific agent
curl -fsSL https://raw.githubusercontent.com/spuentesp/lain/main/install.sh | \
bash /dev/stdin --agent cursor --yes
# See all options
curl -fsSL https://raw.githubusercontent.com/spuentesp/lain/main/install.sh | \
bash /dev/stdin --help
Install options:
| Option | Description | Default |
|---|---|---|
--workspace PATH | Workspace path for LAIN | . |
--transport MODE | MCP transport: stdio, http, both | stdio |
--port PORT | HTTP port for MCP server | 9999 |
--agent AGENT | Target agent: auto, claude, cursor, windsurf, cline | auto |
--embedding-model PATH | Path to ONNX embedding model | - |
--download-model | Download default ONNX model (all-MiniLM-L6-v2.onnx, ~120MB) | - |
-y, --yes | Skip all confirmation prompts | - |
After installation:
# Add to PATH (add to ~/.zshrc, ~/.bashrc, or your shell config)
export PATH="$HOME/.local/lain:$PATH"
# Verify installation
lain --version
# Query the graph
lain query "find Function | limit 5"
Homebrew
brew tap lain-ai/tap
brew install lain
# Initialize
lain init
Pre-built Binary
Download the latest release for your platform from GitHub releases, then:
# Make executable
chmod +x lain
# Run directly
./lain --workspace /path/to/your/project --transport stdio
Build from Source
# Clone the repo
git clone https://github.com/spuentesp/lain.git
cd lain
# Build (requires Rust 1.75+)
cargo build --release
# Binary will be at ./target/release/lain
Quick Start
1. Install LAIN
curl -fsSL https://raw.githubusercontent.com/spuentesp/lain/main/install.sh | bash
2. Initialize for Claude Code (or other agents)
# Auto-detect agent (Claude Code, Cursor, Windsurf, Cline)
lain init
# Or specify agent explicitly
lain init --agent claude
3. Run
# Standard mode (for Claude Code)
lain --workspace /path/to/project --transport stdio
# With HTTP diagnostics (web UI at http://localhost:9999)
lain --workspace /path/to/project --transport both --port 9999
# With semantic search (requires ONNX model)
lain --workspace /path/to/project --embedding-model ~/.local/lain/models/all-MiniLM-L6-v2.onnx
4. Verify
# Check health and LSP status
curl -s -X POST http://localhost:9999/mcp -H "Content-Type: application/json" \
-d '{"jsonrpc":"2.0","method":"tools/call","params":{"name":"get_health","arguments":{}},"id":1}'
# Query the graph directly
lain query "find Function | limit 5"
Key Features
Query Language (query_graph)
JSON-based ops array for flexible graph traversals:
{
"ops": [
{ "op": "find", "type": "Function" },
{ "op": "connect", "edge": "Calls", "depth": { "min": 1, "max": 3 } },
{ "op": "filter", "label": "test" },
{ "op": "semantic_filter", "like": "error handling", "threshold": 0.35 },
{ "op": "limit", "count": 10 }
]
}
Available ops: find, connect, filter, semantic_filter, group, sort, limit
Dependency Intelligence
get_call_chainβ Shortest path between two functionsget_blast_radiusβ Everything affected by a changetrace_dependencyβ What a symbol depends onget_coupling_radarβ Files that change together
Architectural Analysis
find_anchorsβ Most-called, most-stable symbols (architectural pillars)list_entry_pointsβ Findmain(), route handlers, app initializationget_context_depthβ How far from an entry point (abstraction layers)explore_architectureβ High-level tree of modules and files
Search
semantic_searchβ Find code by meaning, not just names (uses local ONNX embeddings)
Code Health
find_dead_codeβ Potentially unreachable code (filters trait defaults, common names)suggest_refactor_targetsβ High-coupling, low-stability nodes
Build Integration
Lain enriches build failures with architectural context:
run_buildβ Build with Rust/Go/JS/Python toolchain error parsingrun_testsβ Tests with error enrichmentrun_clippyβ cargo clippy with context
Requirements
| Requirement | Details |
|---|---|
| Rust | 1.75 or newer |
| Git | Required for co-change analysis |
| ONNX Model | Optional β for semantic search |
Optional: Semantic Search
For semantic_search to work, you need an ONNX embedding model. The easiest way to set this up is using the provided install script:
./scripts/install.sh
Alternatively, you can set it up manually:
# Create model directory
mkdir -p .lain/models
# Download all-MiniLM-L6-v2 (or any compatible model)
# Model produces 384-dim embeddings
curl -L https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2/resolve/main/onnx/model.onnx -o .lain/models/model.onnx
curl -L https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2/resolve/main/tokenizer.json -o .lain/models/tokenizer.json
Set the model path:
export LAIN_EMBEDDING_MODEL=$PWD/.lain/models/model.onnx
# or
./lain --embedding-model ./.lain/models/model.onnx ...
Without the model, semantic_search returns "unavailable" but all other features work.
MCP Transport Modes
| Mode | Command | Use Case |
|---|---|---|
stdio | --transport stdio | Claude Code, MCP clients |
http | --transport http --port 9999 | Web diagnostics dashboard |
both | --transport both --port 9999 | Both stdio + diagnostics |
Troubleshooting
LSP servers not ready?
# Install missing language servers
curl -X POST http://localhost:9999/mcp \
-H "Content-Type: application/json" \
-d '{"jsonrpc":"2.0","method":"tools/call","params":{"name":"install_language_server","arguments":{"language":"rust"}},"id":2}'
Graph stale?
# Sync to current git HEAD
curl -X POST http://localhost:9999/mcp \
-H "Content-Type: application/json" \
-d '{"jsonrpc":"2.0","method":"tools/call","params":{"name":"sync_state","arguments":{}},"id":3}'
View all available tools:
curl -s -X POST http://localhost:9999/mcp \
-H "Content-Type: application/json" \
-d '{"jsonrpc":"2.0","method":"tools/call","params":{"name":"get_agent_strategy","arguments":{}},"id":4}'
A/B Testing Results
A simple A/B test was run on the asciinema_fix_pty_bug (a small fork i made from https://github.com/asciinema/asciinema.git ) across 5 passes, 4 times using a script. Median numbers are reported.
| Metric | with_lain | without_lain |
|---|---|---|
| Pass rate | 5/5 (100%) | 5/5 (100%) |
| Median duration | 39.3s | 54.1s |
| Median tokens in | 35,488 | 41,731 |
Key observations:
- Both conditions passed 100% β the bug fix worked in both conditions, with variation per run.
with_lainused fewer input tokens (~35k vs ~42k median), a difference of ~7k tokens per run.
About the bug: The failing test (pty::tests::spawn_extra_env on macOS) stems from handle_child() setting env vars via env::set_var() before execvp(). The shell's interpretation of echo -n $VAR varies across platforms β sometimes -n is treated as a literal argument. The fix: use printf "%s" "$ASCIINEMA_TEST_FOO" instead, portable across all Unix-like systems.
This was a test I did for A/B comparison β not a rigorous evaluation.
License
MIT β Copyright (c) 2026 spuentesp
