Document OpenCode integration for Docker Model Runner and WSL troubleshooting (#24052)

Added OpenCode integration details and configuration instructions. Also
added a troubleshooting hint for WSL

<!--Delete sections as needed -->

## Description

Added instructions on how to use OpenCode based on the official blog
post for the topic

## Related issues or tickets

<!-- Related issues, pull requests, or Jira tickets -->

## Reviews

<!-- Notes for reviewers here -->
<!-- List applicable reviews (optionally @tag reviewers) -->

- [ ] Technical review
- [ ] Editorial review
- [ ] Product review

---------

Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
This commit is contained in:
Tobias Fenster
2026-02-02 16:24:23 +01:00
committed by GitHub
parent 8e963e52bf
commit 53aa093915

View File

@@ -224,6 +224,40 @@ response = llm.complete("Write a hello world function")
print(response.text)
```
## OpenCode
[OpenCode](https://opencode.ai/) is an open-source coding assistant designed to integrate directly into developer workflows. It supports multiple model providers and exposes a flexible configuration system that makes it easy to switch between them.
### Configuration
1. Install OpenCode (see [docs](https://opencode.ai/docs/#install))
2. Reference DMR in your OpenCode configuration, either globally at `~/.config/opencode/opencode.json` or project specific with a `opencode.json` file in the root of your project
```json
{
"$schema": "https://opencode.ai/config.json",
"provider": {
"dmr": {
"npm": "@ai-sdk/openai-compatible",
"name": "Docker Model Runner",
"options": {
"baseURL": "http://localhost:12434/v1"
},
"models": {
"ai/qwen2.5-coder": {
"name": "ai/qwen2.5-coder"
},
"ai/llama3.2": {
"name": "ai/llama3.2"
}
}
}
}
}
```
3. Select the model you want in OpenCode
You can find more details in [this Docker Blog post](https://www.docker.com/blog/opencode-docker-model-runner-private-ai-coding/)
## Common issues
### "Connection refused" errors
@@ -240,6 +274,8 @@ print(response.text)
3. Check if another service is using port 12434.
4. If you run your tool in WSL and want to connect to DMR on the host via `localhost`, this might not directly work. Configuring WSL to use [mirrored networking](https://learn.microsoft.com/en-us/windows/wsl/networking#mirrored-mode-networking) can solve this.
### "Model not found" errors
1. Verify the model is pulled: