Tested Configuration
| Component | Version / Details |
|---|---|
| Machine | Mac M4 Pro, RAM 24GB |
| LM Studio | 0.4.9 |
| OpenClaw | 2026.4.12 |
| ByteRover CLI | 3.3.0 |
This is an experimental setup. ByteRover and autonomous agents can run with OpenClaw on a Apple RAM 24GB machine, but for production usage we recommend at least an Apple M4 with RAM 48GB.
Step 1 — Download the Models
Search for and download both GGUF files directly from LM Studio’s Discover tab.Download Gemma 4 E4B for OpenClaw
Search for 
unsloth/gemma-4-E4B-it-GGUF and download gemma-4-E4B-it-UD-Q4_K_XL.gguf.
On a 24 GB machine, both models fit in memory simultaneously. Gemma 4 E4B at Q4 uses ~8.7 GB and Qwen3.5-9B at Q4 uses ~10.5 GB.
Step 2 — Load Both Models in LM Studio
LM Studio serves all loaded models from a single endpoint athttp://localhost:1234/v1. Load both models before starting the server.
Load Gemma 4 E4B
Click on 
gemma-4-E4B-it-UD-Q4_K_XL.gguf and click Load. Note the API Identifier — LM Studio assigns it google/gemma-4-e4b. This is the model ID you will use in OpenClaw’s config.
Load Qwen3.5-9B
Click on 
Qwen3.5-9B-Q4_K_S.gguf.gguf and click Load. The API Identifier will be qwen3.5-9b.
Step 3 — Configure Your Agent
Both OpenClaw and Hermes use the same local provider setup. Pick the agent you are using.- OpenClaw
- Hermes
Run the OpenClaw onboard wizard:
Select Custom Provider
When prompted for Model/auth provider, scroll down and select Custom Provider.

Enter the endpoint details
Fill in the following when prompted:
The wizard verifies the endpoint and reports Verification successful.
| Field | Value |
|---|---|
| API Base URL | http://localhost:1234/v1 |
| API Key | (leave blank) |
| Endpoint compatibility | OpenAI-compatible |
| Model ID | google/gemma-4-e4b |
| Model alias | google-gemma-4-e4b |

Resulting openclaw.json config
Resulting openclaw.json config
The wizard writes the following into
~/.openclaw/openclaw.json. You can also add this manually:Step 4 — Configure ByteRover CLI
Connect ByteRover to the same local endpoint and select the Qwen model.- TUI
- CLI
Select OpenAI Compatible
Scroll to OpenAI Compatible and press Enter. This covers LM Studio, Ollama, and any other OpenAI-compatible local server.

Enter the base URL
When prompted, enter 
http://localhost:1234/v1 and press Enter. Leave the API key blank.
Step 5 — Verify ByteRover Is Working
Run a quick curate command to confirm ByteRover is using the local Qwen model.Confirm it processes with the local model
ByteRover sends the request to Qwen3.5-9B on LM Studio. You can watch the LM Studio Developer tab update in real time.

Step 6 — Enable ByteRover Memory Integration
Connect your agent to ByteRover for persistent memory across sessions.OpenClaw Integration
Configure ByteRover as the context engine for OpenClaw
Hermes Integration
Configure ByteRover as the memory provider for Hermes
Reference
LLM Providers
Connect an external provider or use the built-in LLM
Onboard Context
Learn how to seed your context tree with existing knowledge
Reference
Configuration details, troubleshooting, and advanced topics
Local & Cloud
Exploring local & cloud options















