diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index 145d965d..178fc5f6 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -9,40 +9,136 @@ "models": [ { "id": "qwen3-235b:strip_thinking_response=true", - "name": "Venice Large (qwen3-235b)", - "cost_per_1m_in": 1.5, - "cost_per_1m_out": 6, + "name": "Venice Large 1.1 (qwen3-235b)", + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 3.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 50000, + "default_max_tokens": 16384, "can_reason": true, "supports_attachments": false }, { "id": "qwen3-4b:strip_thinking_response=true", "name": "Venice Small (qwen3-4b)", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 32768, - "default_max_tokens": 25000, + "default_max_tokens": 16384, "can_reason": true, "supports_attachments": false }, { "id": "mistral-31-24b", - "name": "Venice Medium (mistral-31-24b)", + "name": "Venice Medium (Mistral 3.1 24B)", "cost_per_1m_in": 0.5, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 50000, + "default_max_tokens": 16384, + "can_reason": false, + "supports_attachments": true + }, + { + "id": "qwen3-235b-a22b-thinking-2507:strip_thinking_response=true", + "name": "Qwen 3 235B A22B Thinking 2507", + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 3.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 16384, + "can_reason": true, + "supports_attachments": false + }, + { + "id": "qwen3-235b-a22b-instruct-2507", + "name": "Qwen 3 235B A22B Instruct 2507", + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.75, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 16384, + "can_reason": false, + "supports_attachments": false + }, + { + "id": "qwen3-next-80b", + "name": "Qwen 3 Next 80b", + "cost_per_1m_in": 0.35, + "cost_per_1m_out": 1.9, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 16384, + "can_reason": false, + "supports_attachments": false + }, + { + "id": "qwen3-coder-480b-a35b-instruct", + "name": "Qwen 3 Coder 480b", + "cost_per_1m_in": 0.75, + "cost_per_1m_out": 3, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 16384, + "can_reason": false, + "supports_attachments": false + }, + { + "id": "google-gemma-3-27b-it", + "name": "Google Gemma 3 27B Instruct", + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 202752, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true }, + { + "id": "openai-gpt-oss-120b", + "name": "OpenAI GPT OSS 120B", + "cost_per_1m_in": 0.07, + "cost_per_1m_out": 0.3, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 16384, + "can_reason": false, + "supports_attachments": false + }, + { + "id": "zai-org-glm-4.6", + "name": "GLM 4.6", + "cost_per_1m_in": 0.85, + "cost_per_1m_out": 2.75, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 202752, + "default_max_tokens": 16384, + "can_reason": false, + "supports_attachments": false + }, + { + "id": "deepseek-ai-DeepSeek-R1", + "name": "DeepSeek R1", + "cost_per_1m_in": 0.85, + "cost_per_1m_out": 2.75, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 16384, + "can_reason": false, + "supports_attachments": false + }, { "id": "llama-3.2-3b", "name": "Llama 3.2 3B", @@ -51,7 +147,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 25000, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false }, @@ -62,8 +158,8 @@ "cost_per_1m_out": 2.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 65536, - "default_max_tokens": 32000, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false }