mirror of
https://github.com/lobehub/lobehub.git
synced 2026-03-27 13:29:15 +07:00
* ✨ feat: add GPT-5.4 model support and fix reasoning payload pruning - Add GPT-5.4 model card to model-bank - Update planCardModels to use gpt-5.4 - Add gpt-5.4 to responsesAPIModels - Fix pruneReasoningPayload to strip logprobs/top_logprobs for reasoning models - Add logprobs, top_logprobs to ChatStreamPayload type - Extend reasoning_effort to include none and xhigh - Add success log for non-fallback requests in RouterRuntime - Fix log parameter mismatch in RouterRuntime Fixes LOBE-5735 * 🐛 fix: match gpt-5.4 to gpt5_2ReasoningEffort in openrouter and vercelaigateway * 🐛 fix: update OpenRouterReasoning effort type to include none and xhigh * 🐛 fix: use tiered pricing for gpt-5.4 based on 272K token threshold * 🌐 chore: update i18n translations * 🐛 fix: update claude-sonnet model version to 4-6 in planCardModels * ✨ feat: add GPT-5.4 Pro model support * 🐛 fix: remove dated snapshot for gpt-5.4-pro in responsesAPIModels * 🐛 fix: add tierBy support for cross-unit tiered pricing threshold OpenAI charges output at 1.5x when INPUT exceeds 272K tokens. The tiered strategy previously only checked the unit's own quantity to select a tier. Added optional tierBy field to TieredPricingUnit so output/cacheRead tiers can reference input quantity for selection. * 🐛 fix: use totalInputTokens for tiered pricing tier selection Tiered pricing tiers should be determined by total prompt size (totalInputTokens), not each unit's own quantity. This fixes output and cacheRead being charged at the wrong tier rate when the prompt exceeds the threshold but the individual unit quantity does not.