diff --git a/langgraphics/metadata/models.json b/langgraphics/metadata/models.json index 684a898..9222cf2 100644 --- a/langgraphics/metadata/models.json +++ b/langgraphics/metadata/models.json @@ -10091,13 +10091,13 @@ "attachment": true, "temperature": true, "limit": { - "output": 262144, + "output": 131000, "context": 262144 }, "cost": { - "input": 0.95, - "output": 4, - "cache_read": 0.16, + "input": 1.2, + "output": 4.5, + "cache_read": 0.2, "cache_write": 0 } }, @@ -12281,6 +12281,22 @@ "cache_write": 0 } }, + "pro/moonshotai/kimi-k2.6": { + "reasoning": true, + "tool_call": true, + "attachment": false, + "temperature": true, + "limit": { + "output": 262000, + "context": 262000 + }, + "cost": { + "input": 0.95, + "output": 4, + "cache_read": 0.16, + "cache_write": 0 + } + }, "pro/moonshotai/kimi-k2-instruct-0905": { "reasoning": false, "tool_call": true, @@ -12307,8 +12323,8 @@ "context": 262000 }, "cost": { - "input": 0.55, - "output": 3, + "input": 0.45, + "output": 2.25, "cache_read": 0, "cache_write": 0 } @@ -16432,6 +16448,22 @@ "cache_write": 0 } }, + "deepseek/deepseek-r1-distill-qwen-32b": { + "reasoning": true, + "tool_call": false, + "attachment": false, + "temperature": true, + "limit": { + "output": 32768, + "context": 32768 + }, + "cost": { + "input": 0.29, + "output": 0.29, + "cache_read": 0, + "cache_write": 0 + } + }, "deepseek/deepseek-v3-turbo": { "reasoning": false, "tool_call": true, @@ -16448,6 +16480,38 @@ "cache_write": 0 } }, + "deepseek/deepseek-r1-distill-qwen-14b": { + "reasoning": true, + "tool_call": false, + "attachment": false, + "temperature": true, + "limit": { + "output": 16384, + "context": 32768 + }, + "cost": { + "input": 0.15, + "output": 0.15, + "cache_read": 0, + "cache_write": 0 + } + }, + "inclusionai/ling-2.6-1t": { + "reasoning": false, + "tool_call": true, + "attachment": false, + "temperature": true, + "limit": { + "output": 32768, + "context": 262144 + }, + "cost": { + "input": 0, + "output": 0, + "cache_read": 0, + "cache_write": 0 + } + }, "nousresearch/hermes-2-pro-llama-3-8b": { "reasoning": false, "tool_call": false, @@ -17441,6 +17505,22 @@ "cache_write": 0 } }, + "moonshotai/kimi-k2.6-tee": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 262144, + "context": 262144 + }, + "cost": { + "input": 0.44, + "output": 2, + "cache_read": 0, + "cache_write": 0 + } + }, "moonshotai/kimi-k2.5-tee": { "reasoning": true, "tool_call": true, @@ -18513,22 +18593,6 @@ "cache_write": 0 } }, - "deepseek/deepseek-r1-distill-qwen-32b": { - "reasoning": true, - "tool_call": false, - "attachment": false, - "temperature": true, - "limit": { - "output": 32768, - "context": 32768 - }, - "cost": { - "input": 0.29, - "output": 0.29, - "cache_read": 0, - "cache_write": 0 - } - }, "openrouter/auto": { "reasoning": true, "tool_call": true, @@ -21673,6 +21737,22 @@ "cache_write": 0 } }, + "hy3-preview-free": { + "reasoning": true, + "tool_call": true, + "attachment": false, + "temperature": true, + "limit": { + "output": 64000, + "context": 256000 + }, + "cost": { + "input": 0, + "output": 0, + "cache_read": 0, + "cache_write": 0 + } + }, "qwen3-coder": { "reasoning": false, "tool_call": true, @@ -22671,6 +22751,22 @@ "cache_write": 0 } }, + "openai/gpt-image-2": { + "reasoning": false, + "tool_call": false, + "attachment": true, + "temperature": false, + "limit": { + "output": 0, + "context": 0 + }, + "cost": { + "input": 5.0505, + "output": 32.3232, + "cache_read": 1.2626, + "cache_write": 0 + } + }, "openai/gpt-4-classic-0314": { "reasoning": false, "tool_call": true, @@ -23929,7 +24025,7 @@ "attachment": true, "temperature": false, "limit": { - "output": 8192, + "output": 262144, "context": 262144 }, "cost": { @@ -29956,6 +30052,23 @@ "cache_write": 0 } }, + "gpt-5.5": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": false, + "limit": { + "output": 130000, + "context": 1050000, + "input": 920000 + }, + "cost": { + "input": 5, + "output": 30, + "cache_read": 0.5, + "cache_write": 0 + } + }, "o4-mini-deep-research": { "reasoning": true, "tool_call": true, @@ -35033,7 +35146,7 @@ "cache_write": 0 } }, - "gemma-4-26b-it": { + "gemma-4-26b-a4b-it": { "reasoning": true, "tool_call": true, "attachment": false,