🤖 docs: Add Truefoundry Custom Endpoint Example (#374)

* added truefoundry mdx

* docs: update TrueFoundry configuration with model naming and structured features

* docs: simplify TrueFoundry documentation by removing Key Features section

* updated tfy docs for librechat

* added truefoundry ai gateway in docs
This commit is contained in:
rishiraj
2025-08-14 01:04:33 +05:30
committed by GitHub
parent 92ceede0da
commit 6b023df4fc

View File

@@ -0,0 +1,31 @@
---
title: TrueFoundry AI Gateway
---
TrueFoundry AI Gateway is the proxy layer that sits between your applications and the LLM providers and MCP Servers. It is an enterprise-grade platform that enables users to access 1000+ LLMs using a unified interface while taking care of observability and governance.
## Configuration Details
To use [TrueFoundry's AI Gateway](https://www.truefoundry.com/ai-gateway) follow the quick start guide [here](https://docs.truefoundry.com/gateway/quick-start).
### Getting Base URL and Model Names
- **Base URL and Model Names**: Get your TrueFoundry AI Gateway endpoint URL and model name from the unified code snippet (ensure you have added the same model name)
- **PAT**: Generate from your TrueFoundry Personal Access Token [PAT](https://docs.truefoundry.com/gateway/authentication#personal-access-token-pat)
## Using TrueFoundry AI Gateway with any LLM model
```yaml filename="librechat.yaml"
- name: "TrueFoundry"
apiKey: "${TRUEFOUNDRY_API_KEY}"
baseURL: "${TRUEFOUNDRY_GATEWAY_URL}"
models:
default: ["openai-main/gpt-4o-mini", "openai-main/gpt-4o"]
fetch: true
titleConvo: true
titleModel: "current_model"
summarize: false
summaryModel: "current_model"
forcePrompt: false
modelDisplayLabel: "TrueFoundry:OpenAI"
```
For more details you can check: [TrueFoundry Docs](https://docs.truefoundry.com/docs/introduction)