LMStudio MCP
A Model Control Protocol (MCP) server that allows Claude to communicate with locally running LLM models via LM Studio.
Ask AI about LMStudio MCP
Powered by Claude Β· Grounded in docs
I know everything about LMStudio MCP. Ask me about installation, configuration, usage, or troubleshooting.
0/500
Reviews
Documentation
LMStudio-MCP
A Model Control Protocol (MCP) server that allows Claude to communicate with locally running LLM models via LM Studio.
Overview
LMStudio-MCP creates a bridge between Claude (with MCP capabilities) and your locally running LM Studio instance. This allows Claude to:
- Check the health of your LM Studio API
- List available models
- Get the currently loaded model
- Generate chat and raw text completions using your local models
- Generate vector embeddings for semantic search and RAG
- Hold stateful multi-turn conversations via response IDs
- Start and continue persistent conversations with a locked-in system prompt
This enables you to leverage your own locally running models through Claude's interface, combining Claude's capabilities with your private models.
Prerequisites
- Python 3.7+
- LM Studio installed and running locally with a model loaded
- Claude with MCP access
- Required Python packages (see Installation)
π Quick Installation
One-Line Install (Recommended)
curl -fsSL https://raw.githubusercontent.com/infinitimeless/LMStudio-MCP/main/install.sh | bash
Manual Installation Methods
1. Local Python Installation
git clone https://github.com/infinitimeless/LMStudio-MCP.git
cd LMStudio-MCP
pip install requests "mcp[cli]" openai
2. Docker Installation
# Using pre-built image
docker run -it --network host ghcr.io/infinitimeless/lmstudio-mcp:latest
# Or build locally
git clone https://github.com/infinitimeless/LMStudio-MCP.git
cd LMStudio-MCP
docker build -t lmstudio-mcp .
docker run -it --network host lmstudio-mcp
3. Docker Compose
git clone https://github.com/infinitimeless/LMStudio-MCP.git
cd LMStudio-MCP
docker-compose up -d
For detailed deployment instructions, see DOCKER.md.
βοΈ Configuration
The bridge supports flexible configuration for different deployment scenarios:
- Default: Connects to
http://localhost:1234/v1 - Custom Host: Set
LMSTUDIO_HOSTenvironment variable (e.g.,192.168.1.100) - Custom Port: Set
LMSTUDIO_PORTenvironment variable (e.g.,5678)
Example:
export LMSTUDIO_HOST=192.168.1.100
export LMSTUDIO_PORT=5678
python lmstudio_bridge.py
π For detailed configuration options, see CONFIGURATION.md
MCP Configuration
Quick Setup
Using GitHub directly (simplest):
{
"lmstudio-mcp": {
"command": "uvx",
"args": [
"https://github.com/infinitimeless/LMStudio-MCP"
]
}
}
Using local installation:
{
"lmstudio-mcp": {
"command": "/bin/bash",
"args": [
"-c",
"cd /path/to/LMStudio-MCP && source venv/bin/activate && python lmstudio_bridge.py"
]
}
}
Using Docker:
{
"lmstudio-mcp-docker": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"--network=host",
"ghcr.io/infinitimeless/lmstudio-mcp:latest"
]
}
}
For complete MCP configuration instructions, see MCP_CONFIGURATION.md.
Optional: MCP description hint
You can add a description field to your .mcp.json entry to help Claude understand when to use this server and what to expect. This is particularly useful for reminding Claude of version requirements:
{
"lmstudio-mcp": {
"command": "...",
"args": [...],
"description": "Local LLM bridge via LM Studio. Use for private/offline inference, embeddings, and multi-turn conversations. start_conversation and continue_conversation require LM Studio v0.3.29+."
}
}
π§ LM Studio System Prompt (Recommended)
Setting a system prompt directly in LM Studio gives your local model a consistent baseline personality and behaviour across all interactions β without needing to pass it on every API call.
How to set it
- Open LM Studio
- Click the model name at the top of the chat panel
- Find the System Prompt field (may be under a βοΈ gear icon or Advanced settings)
- Paste your system prompt and save
The system prompt set here applies to all completions sent via the API, including those from this MCP bridge.
Example system prompts
General assistant β clean and direct:
You are a helpful, concise assistant. Answer directly without preamble like
"Sure!" or "Of course!". Never cut off mid-sentence β always finish your thought.
Casual conversation partner:
You are a regular person having a relaxed conversation with a friend.
Keep responses short and natural, like real chat. No bullet points or formal
language. You can invent fun details about your life and stay consistent with them.
Never cut off mid-sentence β always finish your thought.
Local coding assistant:
You are an expert software engineer. Be concise and precise. When writing code,
always include brief inline comments. Prefer simple, readable solutions over
clever ones. Never cut off mid-sentence or mid-code block.
Privacy-first document analyst:
You are a careful document analyst. Summarise accurately and concisely.
Never invent information not present in the source material.
Always flag uncertainty explicitly.
π‘ Tip: Always end your system prompt with "Never cut off mid-sentence β always finish your thought." This prevents truncated responses regardless of how
max_tokensis configured.
Usage
- Start LM Studio and ensure it's running on port 1234 (the default)
- Set a system prompt in LM Studio (see above β recommended)
- Load a model in LM Studio
- Configure Claude MCP with one of the configurations above
- Connect to the MCP server in Claude when prompted
Available Tools
The bridge provides the following 9 tools:
| Tool | Description |
|---|---|
health_check() | Verify if LM Studio API is accessible |
list_models() | Get a list of all available models in LM Studio |
get_current_model() | Identify which model is currently loaded |
chat_completion(prompt, system_prompt, temperature, max_tokens) | Generate a chat response from your local model |
text_completion(prompt, temperature, max_tokens, stop_sequences) | Generate raw text/code completion β faster, no chat formatting overhead |
generate_embeddings(text, model) | Generate vector embeddings for semantic search and RAG workflows |
create_response(input_text, previous_response_id, reasoning_effort, stream, model) | Stateful conversation via response IDs β requires LM Studio v0.3.29+ |
start_conversation(system_prompt, first_message, temperature, max_tokens, model) | Start a multi-turn session with a persistent system prompt β returns a response_id |
continue_conversation(response_id, message, temperature, max_tokens, model) | Continue a session started with start_conversation β context preserved automatically |
Multi-turn conversation workflow
The recommended way to run a persistent conversation with a local model:
1. start_conversation(
system_prompt="You are a friend at a bar, keep it casual and fun.",
first_message="Hey! How's it going?"
)
β { response_id: "resp_abc...", message: "Hey! Not bad, just unwinding..." }
2. continue_conversation(
response_id="resp_abc...",
message="Work's been insane this week."
)
β { response_id: "resp_def...", message: "Ugh, tell me about it..." }
3. continue_conversation(
response_id="resp_def...",
message="If you could go anywhere tomorrow, where would you go?"
)
β { response_id: "resp_ghi...", message: "Honestly? Northern Portugal..." }
The system prompt is locked in for the entire session β no need to re-send it on every turn. Requires LM Studio v0.3.29+.
Deployment Options
This project supports multiple deployment methods:
| Method | Use Case | Pros | Cons |
|---|---|---|---|
| Local Python | Development, simple setup | Fast, direct control | Requires Python setup |
| Docker | Isolated environments | Clean, portable | Requires Docker |
| Docker Compose | Production deployments | Easy management | More complex setup |
| Kubernetes | Enterprise/scale | Highly scalable | Complex configuration |
| GitHub Direct | Zero setup | No local install needed | Requires internet |
Known Limitations
- Some models (e.g., phi-3.5-mini-instruct_uncensored) may have compatibility issues
- The bridge currently uses only the OpenAI-compatible API endpoints of LM Studio
- Model responses will be limited by the capabilities of your locally loaded model
create_response,start_conversation, andcontinue_conversationrequire LM Studio v0.3.29+generate_embeddingsrequires an embedding-specific model (e.g.text-embedding-nomic-embed-text-v1.5)
Troubleshooting
API Connection Issues
If Claude reports 404 errors when trying to connect to LM Studio:
- Ensure LM Studio is running and has a model loaded
- Check that LM Studio's server is running on port 1234
- Verify your firewall isn't blocking the connection
- Try using "127.0.0.1" instead of "localhost" in the API URL if issues persist
Model Compatibility
If certain models don't work correctly:
- Some models might not fully support the OpenAI chat completions API format
- Try different parameter values (temperature, max_tokens) for problematic models
- Consider switching to a more compatible model if problems persist
For detailed troubleshooting help, see TROUBLESHOOTING.md.
π³ Docker & Containerization
This project includes comprehensive Docker support:
- Multi-architecture images (AMD64, ARM64/Apple Silicon)
- Automated builds via GitHub Actions
- Pre-built images available on GitHub Container Registry
- Docker Compose for easy deployment
- Kubernetes manifests for production deployments
See DOCKER.md for complete containerization documentation.
Contributing
Contributions are welcome! Please see CONTRIBUTING.md for guidelines.
License
MIT
Acknowledgements
This project was originally developed as "Claude-LMStudio-Bridge_V2" and has been renamed and open-sourced as "LMStudio-MCP".
Community Projects
Looking for more advanced features? Check out the community-built enhanced version:
- lmstudio-bridge-enhanced by @ahmedibrahim085 β A powerful extension built on top of this project, adding autonomous agent loops, 37 tools, dynamic MCP discovery, multi-model routing, vision support, and much more.
π If this project helps you, please consider giving it a star!
