- OpenAI-Compatible Client for LM Studio
- What's new
- Install in 5 seconds
- Full text search
- Quick start
- Roadmap & bugs
- Links
Turn LM Studio into a cloud-ready powerhouse. Access local and cloud inference simultaneously in the same app! This fork lets you chat with Cerebras, Groq, OpenRouter, Claude, GPT-4o, Gemini-2.5, DeepSeek, Kimi-K2, GLM V4.5—or any OpenAI-shaped endpoint—without ever leaving the comfy LM Studio UI.
- Native OpenAI sampling knobs (temperature, top-p, frequency penalty, etc.) exposed in the GUI
- Zero config: paste your API key, pick the model, start vibing
- Keeps all local-model superpowers intact—switch between cloud and local on the fly
- Added custom system prompt support
- Navigate to LM Studio Hub
- Hit “Run in LM Studio” on the plugin page
- Done—plugin will be available in Chat view under Your Generators section
Unlock the full power of productivity by instantly navigating between search terms with built-in full-text search across your chat history.
- Grab an API key from your favorite provider (Cerebras, Groq, OpenRouter, Anthropic, OpenAI, etc.)
- In LM Studio → Chat → Your Generators load the plugin → hit Show Settings shortcut → pick Generators tab → paste key
- Type model name, select AI provider, tweak sampling, chat away!
Got feature requests or bugs?
Drop them in the Issues tab—every ticket gets love.
- Plugin Hub page: https://lmstudio.ai/gdmka/openai-compat-endpoint