diff --git a/langgraphics/metadata/models.json b/langgraphics/metadata/models.json index c5d7c88..657d4ae 100644 --- a/langgraphics/metadata/models.json +++ b/langgraphics/metadata/models.json @@ -11068,6 +11068,22 @@ "cache_write": 0 } }, + "zai-glm-5": { + "reasoning": false, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 8192, + "context": 198000 + }, + "cost": { + "input": 1, + "output": 3.2, + "cache_read": 0.2, + "cache_write": 0 + } + }, "deepseek-v3-2": { "reasoning": false, "tool_call": true, @@ -11186,6 +11202,71 @@ "cache_write": 0 } }, + "gemma-3n-e2b-it": { + "reasoning": false, + "tool_call": false, + "attachment": false, + "temperature": true, + "limit": { + "output": 16384, + "context": 1000000 + }, + "cost": { + "input": 0.08, + "output": 0.3, + "cache_read": 0, + "cache_write": 0 + } + }, + "gemma-3-27b-it": { + "reasoning": false, + "tool_call": false, + "attachment": false, + "temperature": false, + "limit": { + "output": 16384, + "context": 32768, + "input": 32768 + }, + "cost": { + "input": 0.306, + "output": 0.306, + "cache_read": 0, + "cache_write": 0 + } + }, + "gemma-3-4b-it": { + "reasoning": false, + "tool_call": false, + "attachment": false, + "temperature": true, + "limit": { + "output": 16384, + "context": 1000000 + }, + "cost": { + "input": 0.08, + "output": 0.3, + "cache_read": 0, + "cache_write": 0 + } + }, + "gemma-3n-e4b-it": { + "reasoning": false, + "tool_call": false, + "attachment": false, + "temperature": true, + "limit": { + "output": 16384, + "context": 1000000 + }, + "cost": { + "input": 0.08, + "output": 0.3, + "cache_read": 0, + "cache_write": 0 + } + }, "gemini-2.5-pro-preview-05-06": { "reasoning": true, "tool_call": false, @@ -11385,6 +11466,22 @@ "cache_write": 0.383 } }, + "gemma-3-12b-it": { + "reasoning": false, + "tool_call": false, + "attachment": false, + "temperature": true, + "limit": { + "output": 16384, + "context": 1000000 + }, + "cost": { + "input": 0.08, + "output": 0.3, + "cache_read": 0, + "cache_write": 0 + } + }, "gemini-live-2.5-flash-preview-native-audio": { "reasoning": true, "tool_call": true, @@ -20812,22 +20909,6 @@ "cache_write": 0 } }, - "gemma-3-12b-it": { - "reasoning": false, - "tool_call": false, - "attachment": false, - "temperature": true, - "limit": { - "output": 16384, - "context": 1000000 - }, - "cost": { - "input": 0.08, - "output": 0.3, - "cache_read": 0, - "cache_write": 0 - } - }, "llama-prompt-guard-2-86m": { "reasoning": false, "tool_call": false, @@ -24998,22 +25079,6 @@ "cache_write": 0 } }, - "gemma-3n-e2b-it": { - "reasoning": false, - "tool_call": false, - "attachment": false, - "temperature": true, - "limit": { - "output": 16384, - "context": 1000000 - }, - "cost": { - "input": 0.08, - "output": 0.3, - "cache_read": 0, - "cache_write": 0 - } - }, "llama-guard-4-12b": { "reasoning": false, "tool_call": false, @@ -25094,22 +25159,6 @@ "cache_write": 0 } }, - "gemma-3-4b-it": { - "reasoning": false, - "tool_call": false, - "attachment": false, - "temperature": true, - "limit": { - "output": 16384, - "context": 1000000 - }, - "cost": { - "input": 0.08, - "output": 0.3, - "cache_read": 0, - "cache_write": 0 - } - }, "ministral-8b-2512": { "reasoning": false, "tool_call": false, @@ -25190,22 +25239,6 @@ "cache_write": 0 } }, - "gemma-3n-e4b-it": { - "reasoning": false, - "tool_call": false, - "attachment": false, - "temperature": true, - "limit": { - "output": 16384, - "context": 1000000 - }, - "cost": { - "input": 0.08, - "output": 0.3, - "cache_read": 0, - "cache_write": 0 - } - }, "codestral-2508": { "reasoning": false, "tool_call": false, @@ -27162,23 +27195,6 @@ "cache_write": 0 } }, - "gemma-3-27b-it": { - "reasoning": false, - "tool_call": false, - "attachment": false, - "temperature": false, - "limit": { - "output": 16384, - "context": 32768, - "input": 32768 - }, - "cost": { - "input": 0.306, - "output": 0.306, - "cache_read": 0, - "cache_write": 0 - } - }, "bge-multilingual-gemma2": { "reasoning": false, "tool_call": false,