From 68ed6d035e44d86ee7f0da348149de27d9f910dc Mon Sep 17 00:00:00 2001 From: GarandPLG Date: Thu, 12 Feb 2026 18:45:10 +0100 Subject: [PATCH] Add Cerebras PAUG provider and split models MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Update default model provider to “Cerebras PAUG”. Reorganize the Cerebras language‑model configuration into two groups – “Cerebras FREE” and “Cerebras PAUG” – each with its own model list and adjusted token limits. --- modules/home/zed/user-settings/agent.nix | 2 +- .../zed/user-settings/language-models.nix | 195 ++++++++++++++---- 2 files changed, 156 insertions(+), 41 deletions(-) diff --git a/modules/home/zed/user-settings/agent.nix b/modules/home/zed/user-settings/agent.nix index 0e8c1a9..78c3aa3 100644 --- a/modules/home/zed/user-settings/agent.nix +++ b/modules/home/zed/user-settings/agent.nix @@ -3,7 +3,7 @@ _: { always_allow_tool_actions = true; default_profile = "ask"; default_model = { - provider = "Cerebras"; + provider = "Cerebras PAUG"; model = "gpt-oss-120b"; }; }; diff --git a/modules/home/zed/user-settings/language-models.nix b/modules/home/zed/user-settings/language-models.nix index 666a11a..3761fc2 100644 --- a/modules/home/zed/user-settings/language-models.nix +++ b/modules/home/zed/user-settings/language-models.nix @@ -1,43 +1,158 @@ _: { - programs.zed-editor.userSettings.language_models.openai_compatible."Cerebras" = { - api_url = "https://api.cerebras.ai/v1"; - available_models = [ - { - name = "qwen-3-235b-a22b-instruct-2507"; - display_name = "Qwen 3 235B Instruct"; - max_tokens = 65000; - max_output_tokens = 32000; - capabilities = { - tools = true; - images = true; - parallel_tool_calls = true; - prompt_cache_key = false; - }; - } - { - name = "llama-3.3-70b"; - display_name = "Llama 3.3 70B"; - max_tokens = 65000; - max_output_tokens = 8000; - capabilities = { - tools = true; - images = true; - parallel_tool_calls = true; - prompt_cache_key = true; - }; - } - { - name = "gpt-oss-120b"; - display_name = "OpenAI GPT OSS"; - max_tokens = 65000; - max_output_tokens = 32000; - capabilities = { - tools = true; - images = true; - parallel_tool_calls = true; - prompt_cache_key = false; - }; - } - ]; + programs.zed-editor.userSettings.language_models.openai_compatible = { + "Cerebras FREE" = { + api_url = "https://api.cerebras.ai/v1"; + available_models = [ + { + name = "gpt-oss-120b"; + display_name = "OpenAI GPT OSS"; + max_tokens = 65000; + max_output_tokens = 32000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = true; + prompt_cache_key = false; + }; + } + { + name = "zai-glm-4.7"; + display_name = "Z.ai GLM 4.7"; + max_tokens = 64000; + max_output_tokens = 40000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = true; + prompt_cache_key = true; + }; + } + { + name = "llama3.1-8b"; + display_name = "Llama 3.1 8B"; + max_tokens = 8000; + max_output_tokens = 8000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = true; + prompt_cache_key = true; + }; + } + { + name = "llama-3.3-70b"; + display_name = "Llama 3.3 70B"; + max_tokens = 65000; + max_output_tokens = 8000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = true; + prompt_cache_key = true; + }; + } + { + name = "qwen-3-32b"; + display_name = "Qwen 3 32B"; + max_tokens = 65000; + max_output_tokens = 8000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = false; + prompt_cache_key = true; + }; + } + { + name = "qwen-3-235b-a22b-instruct-2507"; + display_name = "Qwen 3 235B Instruct"; + max_tokens = 65000; + max_output_tokens = 32000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = false; + prompt_cache_key = true; + }; + } + ]; + }; + "Cerebras PAUG" = { + api_url = "https://api.cerebras.ai/v1"; + available_models = [ + { + name = "gpt-oss-120b"; + display_name = "OpenAI GPT OSS"; + max_tokens = 131000; + max_output_tokens = 40000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = true; + prompt_cache_key = false; + }; + } + { + name = "zai-glm-4.7"; + display_name = "Z.ai GLM 4.7"; + max_tokens = 131000; + max_output_tokens = 40000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = true; + prompt_cache_key = true; + }; + } + { + name = "llama3.1-8b"; + display_name = "Llama 3.1 8B"; + max_tokens = 32000; + max_output_tokens = 8000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = true; + prompt_cache_key = true; + }; + } + { + name = "llama-3.3-70b"; + display_name = "Llama 3.3 70B"; + max_tokens = 128000; + max_output_tokens = 65000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = true; + prompt_cache_key = true; + }; + } + { + name = "qwen-3-32b"; + display_name = "Qwen 3 32B"; + max_tokens = 131000; + max_output_tokens = 8000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = false; + prompt_cache_key = true; + }; + } + { + name = "qwen-3-235b-a22b-instruct-2507"; + display_name = "Qwen 3 235B Instruct"; + max_tokens = 131000; + max_output_tokens = 40000; + capabilities = { + tools = true; + images = true; + parallel_tool_calls = false; + prompt_cache_key = true; + }; + } + ]; + }; }; }