mirror of
https://github.com/ollama/ollama.git
synced 2026-03-27 02:58:43 +07:00
model: add Glm4MoeLiteForCausalLM architecture to support GLM-4.7-Flash (#13779)
This commit is contained in:
@@ -269,6 +269,7 @@ func (kv KV) OllamaEngineRequired() bool {
|
||||
"qwen25vl",
|
||||
"qwen3", "qwen3moe",
|
||||
"qwen3vl", "qwen3vlmoe",
|
||||
"glm4moelite",
|
||||
}, kv.Architecture())
|
||||
}
|
||||
|
||||
@@ -856,6 +857,7 @@ func (f GGML) FlashAttention() bool {
|
||||
return slices.Contains([]string{
|
||||
"bert",
|
||||
"gemma3",
|
||||
"glm4moelite",
|
||||
"gptoss", "gpt-oss",
|
||||
"mistral3",
|
||||
"olmo3",
|
||||
|
||||
Reference in New Issue
Block a user