Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions 4_integrations/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,3 +15,4 @@ We provide guides below for popular tools and are constantly expanding this list

- [Claude Code](/docs/integrations/claude-code)
- [Codex](/docs/integrations/codex)
- [OpenClaw](/docs/integrations/openclaw)
61 changes: 61 additions & 0 deletions 4_integrations/openclaw.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,61 @@
---
title: OpenClaw
description: Use OpenClaw with LM Studio
index: 3
---

OpenClaw now supports LM Studio as a native model provider.
See: [OpenClaw Docs](https://docs.openclaw.ai/providers/lmstudio).

<img src="/assets/marketing/docs/openclaw.webp" style="width: 100%;" data-caption="OpenClaw using LM Studio as a model provider" />

```lms_protip
Have a powerful LLM rig? Use [LM Link](/docs/integrations/lmlink) to run OpenClaw from your laptop while the model runs on your rig.
```

### 1) Start LM Studio's local server

Make sure LM Studio is running as a server (default port `1234`).

You can start it from the app, or from the terminal with `lms`:

```bash
lms server start --port 1234
```

### 2) Run Openclaw with LM Studio as model provider

Install OpenClaw as normal or run the OpenClaw onboard command as follows *(recommended)*

```bash
openclaw onboard
```

and complete the interactive setup with LM Studio as your model provider

You can do the onboarding in non-interactive way by using the following command:

```bash
openclaw onboard \
--non-interactive \
--accept-risk \
--auth-choice lmstudio \
--custom-base-url http://localhost:1234/v1 \
--lmstudio-api-key "$LM_API_TOKEN" \
--custom-model-id qwen/qwen3.5-9b
```

```lms_protip
Use a model (and server/model settings) with more than ~50k context length. Tools like OpenClaw can consume a lot of context.
```

### 3) Set up LM Studio as default memory search provider

To use LM Studio as the embedding model provider for memory search, run the following command and restart openclaw gateway

```bash
openclaw config set agents.defaults.memorySearch.provider lmstudio
openclaw gateway restart
```

If you're running into trouble, hop onto our [Discord](https://discord.gg/lmstudio)
Loading