mirror of
https://github.com/LibreChat-AI/librechat.ai.git
synced 2026-03-27 10:48:32 +07:00
* chore: update GitHub Actions workflow to use latest action version for improved stability * chore: update roadmap references and enhance documentation for AWS Bedrock inference profiles - Updated footer menu and card icons to reflect the 2026 roadmap. - Adjusted default values in changelog and configuration documentation for `maxRecursionLimit`. - Added comprehensive documentation for AWS Bedrock inference profiles, including setup, configuration, and examples. - Introduced Agents API documentation for programmatic access to LibreChat agents. - Enhanced existing documentation for clarity and consistency across various sections. * feat: release Config v1.3.4 with new features and updates - Introduced `create` field in `interface.prompts` and `interface.agents` for enhanced user control. - Added `interface.remoteAgents` configuration for managing remote agent permissions. - Updated `endpoints.bedrock` with `models` and `inferenceProfiles` for better customization. - Included Moonshot as a known endpoint for AI integration. - Introduced new agent capabilities: `deferred_tools` and `programmatic_tools`. - Removed deprecated `forcePrompt` setting from configurations. - Updated default model lists and added support for new models. - Enhanced `modelSpecs` with `artifacts` field and `effort` parameter for Anthropic models. * refactor: update BlogHeader to use usePathname for route handling - Replaced useRouter with usePathname for improved routing in BlogHeader component. - Simplified page retrieval logic by directly using pathname for matching routes. * feat: add changelog for v0.8.3-rc1 release with new features and fixes - Introduced several enhancements including event-driven lazy tool loading, token usage tracking, and programmatic tool calling UI. - Added support for new models and providers, including Claude Opus 4.6 and Moonshot. - Implemented various bug fixes and improvements for better user experience and performance. * chore: nextjs artifact * first draft roadmap * feat: enhance BlogPage with Open Graph image support and layout improvements - Added support for Open Graph images in blog entries to improve visual presentation. - Refactored article layout for better structure and readability, including adjustments to the display of metadata and content. - Updated styles for improved user experience during hover interactions. * feat: enhance BlogPage with date formatting and layout adjustments - Added a new dateFormatted field to entries for improved date display. - Implemented a date formatter for consistent date presentation. - Refactored article layout to use a grid system for better responsiveness. - Updated styles for article links and metadata for enhanced user experience. * feat: add responsive image sizes to BlogPage for improved layout - Included sizes attribute for Open Graph images to enhance responsiveness on different screen sizes. * feat: update feature titles and descriptions for clarity - Changed titles for "Forking Messages and Conversations" to "Forking Chats" and "Memory" to "User Memory" for better alignment with functionality. - Updated descriptions for "Message Search" and "Upload as Text" to enhance understanding of features. * chore: update configuration version to 1.3.4 across multiple documentation files - Updated the version number in `librechat.yaml` examples to reflect the latest release (1.3.4) in various configuration and feature documentation files. * feat: enhance User Memory documentation for clarity and detail - Updated the description to clarify that User Memory is a key/value store that operates on every chat request. - Added a callout to distinguish between key/value storage and conversation memory. - Expanded on the functionality of the memory agent, including its execution process and user control features. - Introduced a section on future improvements for the memory agent's efficiency and relevance. * feat: update title and description for NGINX documentation - Changed the title from "Secure Deployment with Nginx" to "NGINX" for brevity. - Updated the description to provide a clearer overview of the guide's purpose in securing LibreChat deployment with Nginx as a reverse proxy and HTTPS. * feat: update 2026 roadmap with key accomplishments and future plans - Celebrated LibreChat's 3rd anniversary with a summary of achievements from 2025, including growth in GitHub stars and community engagement. - Clarified the timeline for open-sourcing the Code Interpreter API by the end of Q1. - Revised notes on the v1 Admin Panel's core capabilities and community-driven items for better clarity and detail. * feat: enhance blog and author components with Open Graph image support - Added optional `ogImagePosition` field to blog entries for better image placement control. - Updated BlogPage and individual post pages to utilize the new `ogImagePosition` for responsive image styling. - Improved Author component to conditionally render author images based on availability. - Updated 2026 roadmap blog post with a new Open Graph image and position for enhanced visual appeal. * feat: enhance CardComponent with icon support and layout improvements - Added optional `icon` prop to CardComponent for better visual representation. - Updated CardComponent layout to include icon alongside title and children. - Improved styling for CardComponent and CardsBase for enhanced responsiveness and user experience. * feat: update 2026 roadmap with detailed focus areas and community-driven items - Added sections for Q1 and Q2 focus areas, outlining major initiatives like Dynamic Context and Admin Panel. - Enhanced clarity on community-driven items and their prioritization based on GitHub reactions. - Included hiring information to attract full-stack developers for ongoing project support. - Improved overall structure and readability of the roadmap content. * fix: improve icon styling in CardCompat component for better responsiveness - Updated icon container styling to ensure consistent height and width for SVG icons. - Enhanced layout of CardCompat to maintain visual integrity across different screen sizes. * chore: update .gitignore to include next-env.d.ts for TypeScript support * fix: correct import statement formatting in next-env.d.ts for consistency * fix: refine wording in 2026 roadmap for clarity - Updated the description of agentic workflows to emphasize a lean approach to context pulling. - Enhanced overall readability of the section on Dynamic Context. * feat: expand Admin Panel section in 2026 roadmap with detailed capabilities - Added comprehensive descriptions of the Admin Panel's core functionalities, including GUI for configuration, configuration profiles, group and role management, and access controls. - Clarified the development approach for the Admin Panel, emphasizing ongoing iteration and community involvement. - Updated note on the Admin Panel's prioritization and requirements following the ClickHouse acquisition. * feat: add TrackedLink component for enhanced analytics tracking - Introduced a new TrackedLink component that integrates Vercel analytics to track user interactions with links. - The component allows for customizable link properties while ensuring tracking of clicks with relevant metadata. - Updated CardCompat to utilize the new TrackedLink for improved user engagement tracking. * feat: enhance blog post layout and introduce TrackedAnchor component for link tracking - Wrapped the InlineTOC component in a div for improved spacing in blog posts. - Added a new TrackedAnchor component to facilitate link tracking with Vercel analytics, allowing for customizable anchor elements. - Updated mdx-components to utilize TrackedAnchor for enhanced link interaction tracking. * feat: update TrackedLink and TrackedAnchor components for external link handling - Enhanced the TrackedLink component to differentiate between internal and external links, using Next.js Link for internal navigation. - Introduced a utility function to determine if a link is external, improving tracking accuracy. - Updated TrackedAnchor to utilize the same external link handling logic for consistency in link tracking. * feat: add uncaught exception handling section to dotenv configuration documentation - Introduced a new section on uncaught exception handling, explaining how to override the default behavior to keep the app running after exceptions. - Added an option table detailing the `CONTINUE_ON_UNCAUGHT_EXCEPTION` configuration. - Included a warning callout advising against using this feature in production environments. * feat: add ESLint rule for unused variables in TypeScript - Introduced a new ESLint rule to enforce the handling of unused variables, allowing for specific patterns to be ignored. - This enhancement aims to improve code quality by ensuring that developers are alerted to potentially unnecessary variables while maintaining flexibility in naming conventions. * fix: update copyright year in LICENSE file to 2026 * feat: update footer menu link and add 2026 roadmap blog post - Changed the roadmap link in the FooterMenu component to point to the new blog post. - Introduced a new blog post detailing the 2026 roadmap for LibreChat, outlining key features and focus areas for the upcoming year. - Updated the import statement in next-env.d.ts for consistency with the new types directory. * fix: update import path in next-env.d.ts and add comment block in agents.mdx - Changed the import statement in next-env.d.ts to reference the new development types directory. - Added a comment block in agents.mdx to indicate that the Programmatic Tool Calling feature is in private beta. * fix: remove unused ESLint disable comment in context.tsx * chore: update blog
239 lines
9.6 KiB
Plaintext
239 lines
9.6 KiB
Plaintext
---
|
|
title: User Memory
|
|
icon: Bookmark
|
|
description: Key/value store for user memory that runs on every chat request in LibreChat
|
|
---
|
|
|
|
## Overview
|
|
|
|
User Memory in LibreChat is a **key/value store** that persists user-specific information across conversations. A dedicated memory agent runs at the start of **every chat request**, reading from and writing to this store to provide personalized context to the main AI response.
|
|
|
|
<Callout type="info" title="Key/Value Store, Not Conversation Memory">
|
|
This is **not** semantic memory over your entire conversation history. It does not index, embed, or search past conversations. Instead, it maintains a structured set of key/value pairs (e.g., `user_preferences`, `learned_facts`) that are injected into each request as context. Think of it as a persistent notepad the AI reads before every response.
|
|
|
|
For context about previous messages within a single conversation, LibreChat already uses the standard message history window — that is separate from this feature.
|
|
</Callout>
|
|
|
|
<Callout type="important" title="⚠️ Configuration Required">
|
|
Memory functionality must be explicitly configured in your `librechat.yaml` file to work. It is not enabled by default.
|
|
</Callout>
|
|
|
|
## Key Features
|
|
|
|
- **Runs Every Request**: The memory agent executes at the start of each chat request, ensuring stored context is always available
|
|
- **Key/Value Storage**: Information is stored as structured key/value pairs, not as raw conversation logs
|
|
- **Manual Entries**: Users can manually add, edit, or remove memory entries directly, giving full control over what the AI remembers
|
|
- **User Control**: When enabled, users can toggle memory on/off for their individual chats
|
|
- **Customizable Keys**: Restrict what categories of information can be stored using `validKeys`
|
|
- **Token Management**: Set limits on memory usage to control costs
|
|
- **Agent Integration**: Use AI agents to intelligently manage what gets remembered
|
|
|
|
## Configuration
|
|
|
|
To enable memory features, you need to add the `memory` configuration to your `librechat.yaml` file:
|
|
|
|
```yaml filename="librechat.yaml"
|
|
version: 1.3.4
|
|
cache: true
|
|
|
|
memory:
|
|
disabled: false # Set to true to completely disable memory
|
|
personalize: true # Gives users the ability to toggle memory on/off, true by default
|
|
tokenLimit: 2000 # Maximum tokens for memory storage
|
|
messageWindowSize: 5 # Number of recent messages to consider
|
|
agent:
|
|
provider: "openAI"
|
|
model: "gpt-4"
|
|
```
|
|
|
|
The provider field should match the accepted values as defined in the [Model Spec Guide](/docs/configuration/librechat_yaml/object_structure/model_specs#endpoint).
|
|
|
|
**Note:** If you are using a custom endpoint, the endpoint value must match the defined custom endpoint name exactly.
|
|
|
|
See the [Memory Configuration Guide](/docs/configuration/librechat_yaml/object_structure/memory) for detailed configuration options.
|
|
|
|
## How It Works
|
|
|
|
<Callout type="note" title="Memory Agent Execution">
|
|
The memory agent runs on **every chat request** when memory is enabled. It executes concurrently with the main chat response — it begins before the main response starts and is limited to the duration of the main request plus up to 3 seconds after it finishes.
|
|
|
|
This means every message you send triggers the memory agent to:
|
|
1. **Read** the current key/value store and inject relevant entries as context
|
|
2. **Analyze** the recent message window for information worth storing or updating
|
|
3. **Write** any new or modified entries back to the store
|
|
</Callout>
|
|
|
|
### 1. Key/Value Storage
|
|
Memory entries are stored as key/value pairs. When memory is enabled, the system can store entries such as:
|
|
- User preferences (communication style, topics of interest)
|
|
- Important facts explicitly shared by users
|
|
- Ongoing projects or tasks mentioned
|
|
- Any category you define via `validKeys`
|
|
|
|
Users can also **manually create, edit, and delete** memory entries through the interface, giving direct control over what the AI knows about them.
|
|
|
|
### 2. Context Window
|
|
The `messageWindowSize` parameter determines how many recent messages are analyzed for memory updates. This helps the memory agent decide what information is worth storing or updating in the key/value store.
|
|
|
|
### 3. User Control
|
|
When `personalize` is set to `true`:
|
|
- Users see a memory toggle in their chat interface
|
|
- They can enable/disable memory for individual conversations
|
|
- Memory settings persist across sessions
|
|
|
|
### 4. Valid Keys
|
|
You can restrict what categories of information are stored by specifying `validKeys`:
|
|
|
|
```yaml filename="memory / validKeys"
|
|
memory:
|
|
validKeys:
|
|
- "user_preferences"
|
|
- "conversation_context"
|
|
- "learned_facts"
|
|
- "personal_information"
|
|
```
|
|
|
|
## Best Practices
|
|
|
|
### 1. Token Limits
|
|
Set appropriate token limits to balance functionality with cost:
|
|
- Higher limits allow more comprehensive memory
|
|
- Lower limits reduce processing costs
|
|
- Consider your usage patterns and budget
|
|
|
|
### 2. Custom Instructions
|
|
When using `validKeys`, provide custom instructions to the memory agent:
|
|
|
|
```yaml filename="memory / agent with instructions"
|
|
memory:
|
|
agent:
|
|
provider: "openAI"
|
|
model: "gpt-4"
|
|
instructions: |
|
|
Store information only in the specified validKeys categories.
|
|
Focus on explicitly stated preferences and important facts.
|
|
Delete outdated or corrected information promptly.
|
|
```
|
|
|
|
### 3. Privacy Considerations
|
|
- Memory stores user information across conversations
|
|
- Ensure users understand what information is being stored
|
|
- Consider implementing data retention policies
|
|
- Provide clear documentation about memory usage
|
|
|
|
## Examples
|
|
|
|
### Basic Configuration
|
|
Enable memory with default settings:
|
|
|
|
```yaml filename="Basic memory config"
|
|
memory:
|
|
tokenLimit: 2000
|
|
agent:
|
|
provider: "openAI"
|
|
model: "gpt-4.1-mini"
|
|
```
|
|
|
|
### Advanced Configuration
|
|
Full configuration with all options:
|
|
|
|
```yaml filename="Advanced memory config"
|
|
memory:
|
|
disabled: false
|
|
validKeys: ["preferences", "context", "facts"]
|
|
tokenLimit: 3000
|
|
personalize: true
|
|
messageWindowSize: 10
|
|
agent:
|
|
provider: "anthropic"
|
|
model: "claude-3-opus-20240229"
|
|
instructions: "Remember only explicitly stated preferences and key facts."
|
|
model_parameters:
|
|
temperature: 0.3
|
|
```
|
|
|
|
For valid model parameters per provider, see the [Model Spec Preset Fields](/docs/configuration/librechat_yaml/object_structure/model_specs#preset-fields).
|
|
|
|
### Using Predefined Agents
|
|
Reference an existing agent by ID:
|
|
|
|
```yaml filename="Memory with agent ID"
|
|
memory:
|
|
agent:
|
|
id: "memory-specialist-001"
|
|
```
|
|
|
|
### Custom Endpoints with Memory
|
|
|
|
Memory fully supports custom endpoints, including those with custom headers and environment variables. When using a custom endpoint, header placeholders and environment variables are properly resolved during memory processing.
|
|
|
|
```yaml filename="librechat.yaml with custom endpoint for memory"
|
|
|
|
endpoints:
|
|
custom:
|
|
- name: 'Custom Memory Endpoint'
|
|
apiKey: 'dummy'
|
|
baseURL: 'https://api.gateway.ai/v1'
|
|
headers:
|
|
x-gateway-api-key: '${GATEWAY_API_KEY}'
|
|
x-gateway-virtual-key: '${GATEWAY_OPENAI_VIRTUAL_KEY}'
|
|
X-User-Identifier: '{{LIBRECHAT_USER_EMAIL}}'
|
|
X-Application-Identifier: 'LibreChat - Test'
|
|
api-key: '${TEST_CUSTOM_API_KEY}'
|
|
models:
|
|
default:
|
|
- 'gpt-4o-mini'
|
|
- 'gpt-4o'
|
|
fetch: false
|
|
|
|
memory:
|
|
disabled: false
|
|
tokenLimit: 3000
|
|
personalize: true
|
|
messageWindowSize: 10
|
|
agent:
|
|
provider: 'Custom Memory Endpoint'
|
|
model: 'gpt-4o-mini'
|
|
```
|
|
|
|
- All [custom endpoint headers](/docs/configuration/librechat_yaml/object_structure/custom_endpoint#headers) are supported
|
|
|
|
## Troubleshooting
|
|
|
|
### Memory Not Working
|
|
1. Verify memory is configured in `librechat.yaml`
|
|
2. Check that `disabled` is set to `false`
|
|
3. Ensure the configured agent/model is available
|
|
4. Verify users have enabled memory in their chat interface
|
|
5. For custom endpoints: ensure the `provider` name matches the custom endpoint `name` exactly
|
|
|
|
### High Token Usage
|
|
1. Reduce `tokenLimit` to control costs
|
|
2. Decrease `messageWindowSize` to analyze fewer messages
|
|
3. Use `validKeys` to restrict what gets stored
|
|
4. Review and optimize agent instructions
|
|
|
|
### Inconsistent Memory
|
|
1. Check if users are toggling memory on/off
|
|
2. Verify token limits aren't being exceeded
|
|
3. Ensure consistent agent configuration
|
|
4. Review stored memory for conflicts
|
|
|
|
### Custom Endpoint Authentication Issues
|
|
1. Verify environment variables are set correctly in your `.env` file
|
|
2. Ensure custom headers use the correct syntax (`${ENV_VAR}` for environment variables, `{{LIBRECHAT_USER_*}}` for user placeholders)
|
|
3. Check that the custom endpoint is working for regular chat completions before testing with memory
|
|
4. Review server logs for authentication errors from the custom endpoint API
|
|
|
|
## Future Improvements
|
|
|
|
The current implementation runs the memory agent on every chat request unconditionally. Planned improvements include:
|
|
|
|
- **Semantic Trigger for Writes**: Detect when a user has explicitly asked the model to remember something (e.g., "Remember that I prefer Python") and only run the memory write agent in those cases, reducing unnecessary processing on routine messages.
|
|
- **Vector Similarity Recall**: Instead of injecting all stored memory entries into every request, use vector embeddings to retrieve only the entries most relevant to the current conversation context, improving both efficiency and relevance.
|
|
|
|
## Related Features
|
|
|
|
- [Agents](/docs/features/agents) - Build custom AI assistants
|
|
- [Presets](/docs/user_guides/presets) - Save conversation settings
|
|
- [Fork Messages](/docs/features/fork) - Branch conversations while maintaining context |