mirror of
https://github.com/open-webui/docs.git
synced 2026-01-02 09:49:38 +07:00
Update rag.md
This commit is contained in:
@@ -10,7 +10,7 @@ If you're using **Ollama**, note that it **defaults to a 2048-token context leng
|
||||
**Why This Is Critical for Web Search:**
|
||||
Web pages typically contain 4,000-8,000+ tokens even after content extraction, including main content, navigation elements, headers, footers, and metadata. With only 2048 tokens available, you're getting less than half the page content, often missing the most relevant information. Even 4096 tokens is frequently insufficient for comprehensive web content analysis.
|
||||
|
||||
**To Fix This:** Navigate to **Admin Panel > Models > Settings** (of your Ollama model) > **Advanced Parameters** and **increase the context length to 8192+ tokens**. This setting specifically applies to Ollama models. For OpenAI and other integrated models, ensure you're using a model with sufficient built-in context length (e.g., GPT-4 Turbo with 128k tokens).
|
||||
**To Fix This:** Navigate to **Admin Panel > Models > Settings** (of your Ollama model) > **Advanced Parameters** and **increase the context length to 8192+ (or rather, more than 16000) tokens**. This setting specifically applies to Ollama models. For OpenAI and other integrated models, ensure you're using a model with sufficient built-in context length (e.g., GPT-4 Turbo with 128k tokens).
|
||||
|
||||
:::
|
||||
|
||||
|
||||
Reference in New Issue
Block a user