|
|
|
|
@@ -2,19 +2,11 @@ version: 0.30.0
|
|
|
|
|
list:
|
|
|
|
|
- provider: openai
|
|
|
|
|
models:
|
|
|
|
|
- name: gpt-5.1
|
|
|
|
|
- name: gpt-5.2
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
input_price: 1.25
|
|
|
|
|
output_price: 10.0
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-5.1-chat-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
input_price: 1.25
|
|
|
|
|
output_price: 10.0
|
|
|
|
|
input_price: 1.75
|
|
|
|
|
output_price: 14.0
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
@@ -26,14 +18,6 @@ list:
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-5-chat-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
input_price: 1.25
|
|
|
|
|
output_price: 10.0
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-5-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
@@ -58,22 +42,6 @@ list:
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-4.1-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1047576
|
|
|
|
|
input_price: 0.4
|
|
|
|
|
output_price: 1.6
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-4.1-nano
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1047576
|
|
|
|
|
input_price: 0.1
|
|
|
|
|
output_price: 0.4
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-4o
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
@@ -82,98 +50,6 @@ list:
|
|
|
|
|
max_output_tokens: 16384
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-4o-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
input_price: 0.15
|
|
|
|
|
output_price: 0.6
|
|
|
|
|
max_output_tokens: 16384
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: o4-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 1.1
|
|
|
|
|
output_price: 4.4
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o4-mini-high
|
|
|
|
|
type: chat
|
|
|
|
|
real_name: o4-mini
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 1.1
|
|
|
|
|
output_price: 4.4
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
reasoning_effort: high
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o3
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 2.0
|
|
|
|
|
output_price: 8.0
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o3-high
|
|
|
|
|
type: chat
|
|
|
|
|
real_name: o3
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 2.0
|
|
|
|
|
output_price: 8.0
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
reasoning_effort: high
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o3-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 1.1
|
|
|
|
|
output_price: 4.4
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o3-mini-high
|
|
|
|
|
type: chat
|
|
|
|
|
real_name: o3-mini
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 1.1
|
|
|
|
|
output_price: 4.4
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
reasoning_effort: high
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: gpt-4-turbo
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
@@ -203,22 +79,6 @@ list:
|
|
|
|
|
max_batch_size: 100
|
|
|
|
|
- provider: gemini
|
|
|
|
|
models:
|
|
|
|
|
- name: gemini-3-flash-preview
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
input_price: 0.0
|
|
|
|
|
output_price: 0.0
|
|
|
|
|
max_output_tokens: 65536
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gemini-3-pro-preview
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
input_price: 0.0
|
|
|
|
|
output_price: 0.0
|
|
|
|
|
max_output_tokens: 65536
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gemini-2.5-flash
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
@@ -243,6 +103,16 @@ list:
|
|
|
|
|
max_output_tokens: 64000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gemini-3-pro-preview
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gemini-3-flash-preview
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gemini-2.0-flash
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
@@ -439,6 +309,13 @@ list:
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- provider: mistral
|
|
|
|
|
models:
|
|
|
|
|
- name: mistral-large-latest
|
|
|
|
|
type: chat
|
|
|
|
|
input_price: 0.5
|
|
|
|
|
output_price: 1.5
|
|
|
|
|
max_output_tokens: 262144
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: mistral-medium-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
@@ -455,32 +332,38 @@ list:
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: magistral-medium-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 40960
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 2.0
|
|
|
|
|
output_price: 5.0
|
|
|
|
|
- name: magistral-small-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 40960
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.5
|
|
|
|
|
output_price: 1.5
|
|
|
|
|
- name: devstral-medium-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 256000
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
input_price: 0.4
|
|
|
|
|
output_price: 2.0
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: devstral-small-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 256000
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
input_price: 0.1
|
|
|
|
|
output_price: 0.3
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: codestral-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 256000
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
input_price: 0.3
|
|
|
|
|
output_price: 0.9
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: ministral-14b-latest
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
input_price: 0.2
|
|
|
|
|
output_price: 0.2
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: mistral-embed
|
|
|
|
|
type: embedding
|
|
|
|
|
max_input_tokens: 8092
|
|
|
|
|
@@ -558,42 +441,24 @@ list:
|
|
|
|
|
max_input_tokens: 4096
|
|
|
|
|
- provider: xai
|
|
|
|
|
models:
|
|
|
|
|
- name: grok-4
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 256000
|
|
|
|
|
input_price: 3.0
|
|
|
|
|
output_price: 15.0
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: grok-4-fast-non-reasoning
|
|
|
|
|
- name: grok-4-1-fast-non-reasoning
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 2000000
|
|
|
|
|
input_price: 0.2
|
|
|
|
|
output_price: 0.5
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: grok-4-fast-reasoning
|
|
|
|
|
- name: grok-4-1-fast-reasoning
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 2000000
|
|
|
|
|
input_price: 0.2
|
|
|
|
|
output_price: 0.5
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: grok-code-fast
|
|
|
|
|
- name: grok-code-fast-1
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 256000
|
|
|
|
|
input_price: 0.2
|
|
|
|
|
output_price: 1.5
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: grok-3
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 3.0
|
|
|
|
|
output_price: 15.0
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: grok-3-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.3
|
|
|
|
|
output_price: 0.5
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- provider: perplexity
|
|
|
|
|
models:
|
|
|
|
|
- name: sonar-pro
|
|
|
|
|
@@ -611,11 +476,6 @@ list:
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
input_price: 2.0
|
|
|
|
|
output_price: 8.0
|
|
|
|
|
- name: sonar-reasoning
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
input_price: 1.0
|
|
|
|
|
output_price: 5.0
|
|
|
|
|
- name: sonar-deep-research
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
@@ -702,6 +562,16 @@ list:
|
|
|
|
|
max_output_tokens: 65536
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gemini-3-pro-preview
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gemini-3-flash-preview
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gemini-2.0-flash-001
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1048576
|
|
|
|
|
@@ -877,18 +747,6 @@ list:
|
|
|
|
|
require_max_tokens: true
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: mistral-small-2503
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 32000
|
|
|
|
|
input_price: 0.1
|
|
|
|
|
output_price: 0.3
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: codestral-2501
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 256000
|
|
|
|
|
input_price: 0.3
|
|
|
|
|
output_price: 0.9
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: text-embedding-005
|
|
|
|
|
type: embedding
|
|
|
|
|
max_input_tokens: 20000
|
|
|
|
|
@@ -1336,32 +1194,23 @@ list:
|
|
|
|
|
max_batch_size: 10
|
|
|
|
|
- provider: hunyuan
|
|
|
|
|
models:
|
|
|
|
|
- name: hunyuan-turbos-latest
|
|
|
|
|
- name: hunyuan-2.0-instruct-20251111
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 28000
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.112
|
|
|
|
|
output_price: 0.28
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: hunyuan-t1-latest
|
|
|
|
|
- name: hunyuan-2.0-thinking-20251109
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 28000
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.14
|
|
|
|
|
output_price: 0.56
|
|
|
|
|
- name: hunyuan-lite
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 250000
|
|
|
|
|
input_price: 0.0
|
|
|
|
|
output_price: 0.0
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: hunyuan-turbos-vision
|
|
|
|
|
- name: hunyuan-vision-1.5-instruct
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 6144
|
|
|
|
|
max_input_tokens: 24576
|
|
|
|
|
input_price: 0.42
|
|
|
|
|
output_price: 0.84
|
|
|
|
|
supports_vision: true
|
|
|
|
|
- name: hunyuan-t1-vision
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 24000
|
|
|
|
|
output_price: 1.26
|
|
|
|
|
supports_vision: true
|
|
|
|
|
- name: hunyuan-embedding
|
|
|
|
|
type: embedding
|
|
|
|
|
@@ -1414,65 +1263,29 @@ list:
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
- provider: zhipuai
|
|
|
|
|
models:
|
|
|
|
|
- name: glm-4.6
|
|
|
|
|
- name: glm-4.7
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 202752
|
|
|
|
|
input_price: 0.28
|
|
|
|
|
output_price: 1.12
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: glm-4.5
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.28
|
|
|
|
|
output_price: 1.12
|
|
|
|
|
- name: glm-4.5-x
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 1.12
|
|
|
|
|
output_price: 4.48
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: glm-4.5-air
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.084
|
|
|
|
|
output_price: 0.56
|
|
|
|
|
- name: glm-4.5-airx
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.56
|
|
|
|
|
output_price: 2.24
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: glm-4.5-flash
|
|
|
|
|
- name: glm-4.7:instruct
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.0
|
|
|
|
|
output_price: 0.0
|
|
|
|
|
- name: glm-4.5v
|
|
|
|
|
real_name: glm-4.7
|
|
|
|
|
max_input_tokens: 202752
|
|
|
|
|
input_price: 0.56
|
|
|
|
|
output_price: 2.24
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
thinking:
|
|
|
|
|
type: disabled
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: glm-4.6v
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 65536
|
|
|
|
|
input_price: 0.56
|
|
|
|
|
output_price: 1.68
|
|
|
|
|
input_price: 0.28
|
|
|
|
|
output_price: 0.84
|
|
|
|
|
supports_vision: true
|
|
|
|
|
- name: glm-z1-air
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.07
|
|
|
|
|
output_price: 0.07
|
|
|
|
|
- name: glm-z1-airx
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.7
|
|
|
|
|
output_price: 0.7
|
|
|
|
|
- name: glm-z1-flashx
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.014
|
|
|
|
|
output_price: 0.014
|
|
|
|
|
- name: glm-z1-flash
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.0
|
|
|
|
|
output_price: 0.0
|
|
|
|
|
- name: embedding-3
|
|
|
|
|
type: embedding
|
|
|
|
|
max_input_tokens: 8192
|
|
|
|
|
@@ -1485,27 +1298,25 @@ list:
|
|
|
|
|
input_price: 0.112
|
|
|
|
|
- provider: minimax
|
|
|
|
|
models:
|
|
|
|
|
- name: minimax-m2
|
|
|
|
|
- name: minimax-m2.1
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 204800
|
|
|
|
|
input_price: 0.294
|
|
|
|
|
output_price: 1.176
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: minimax-m2.1-lightning
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 204800
|
|
|
|
|
input_price: 0.294
|
|
|
|
|
output_price: 2.352
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- provider: openrouter
|
|
|
|
|
models:
|
|
|
|
|
- name: openai/gpt-5.1
|
|
|
|
|
- name: openai/gpt-5.2
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
input_price: 1.25
|
|
|
|
|
output_price: 10.0
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: openai/gpt-5.1-chat
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
input_price: 1.25
|
|
|
|
|
output_price: 10.0
|
|
|
|
|
input_price: 1.75
|
|
|
|
|
output_price: 14.0
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
@@ -1517,14 +1328,6 @@ list:
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: openai/gpt-5-chat
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
input_price: 1.25
|
|
|
|
|
output_price: 10.0
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: openai/gpt-5-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
@@ -1549,22 +1352,6 @@ list:
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: openai/gpt-4.1-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1047576
|
|
|
|
|
input_price: 0.4
|
|
|
|
|
output_price: 1.6
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: openai/gpt-4.1-nano
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1047576
|
|
|
|
|
input_price: 0.1
|
|
|
|
|
output_price: 0.4
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: openai/gpt-4o
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
@@ -1572,91 +1359,6 @@ list:
|
|
|
|
|
output_price: 10.0
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: openai/gpt-4o-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
input_price: 0.15
|
|
|
|
|
output_price: 0.6
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: openai/o4-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 1.1
|
|
|
|
|
output_price: 4.4
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: openai/o4-mini-high
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 1.1
|
|
|
|
|
output_price: 4.4
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
reasoning_effort: high
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: openai/o3
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 2.0
|
|
|
|
|
output_price: 8.0
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: openai/o3-high
|
|
|
|
|
type: chat
|
|
|
|
|
real_name: openai/o3
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 2.0
|
|
|
|
|
output_price: 8.0
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
reasoning_effort: high
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: openai/o3-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 1.1
|
|
|
|
|
output_price: 4.4
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: openai/o3-mini-high
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
input_price: 1.1
|
|
|
|
|
output_price: 4.4
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: openai/gpt-oss-120b
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
@@ -1790,6 +1492,12 @@ list:
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.12
|
|
|
|
|
output_price: 0.3
|
|
|
|
|
- name: mistralai/mistral-large-2512
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
input_price: 0.5
|
|
|
|
|
output_price: 1.5
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: mistralai/mistral-medium-3.1
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
@@ -1803,26 +1511,11 @@ list:
|
|
|
|
|
input_price: 0.1
|
|
|
|
|
output_price: 0.3
|
|
|
|
|
supports_vision: true
|
|
|
|
|
- name: mistralai/magistral-medium-2506
|
|
|
|
|
- name: mistralai/devstral-2512
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 40960
|
|
|
|
|
input_price: 2.0
|
|
|
|
|
output_price: 5.0
|
|
|
|
|
- name: mistralai/magistral-medium-2506:thinking
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 40960
|
|
|
|
|
input_price: 2.0
|
|
|
|
|
output_price: 5.0
|
|
|
|
|
- name: mistralai/magistral-small-2506
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 40960
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
input_price: 0.5
|
|
|
|
|
output_price: 1.5
|
|
|
|
|
- name: mistralai/devstral-medium
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.4
|
|
|
|
|
output_price: 2.0
|
|
|
|
|
output_price: 0.22
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: mistralai/devstral-small
|
|
|
|
|
type: chat
|
|
|
|
|
@@ -1836,6 +1529,12 @@ list:
|
|
|
|
|
input_price: 0.3
|
|
|
|
|
output_price: 0.9
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: mistralai/ministral-14b-2512
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
input_price: 0.2
|
|
|
|
|
output_price: 0.2
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: ai21/jamba-large-1.7
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 256000
|
|
|
|
|
@@ -1860,29 +1559,11 @@ list:
|
|
|
|
|
input_price: 0.0375
|
|
|
|
|
output_price: 0.15
|
|
|
|
|
max_output_tokens: 4096
|
|
|
|
|
- name: deepseek/deepseek-v3.2-exp
|
|
|
|
|
- name: deepseek/deepseek-v3.2
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 163840
|
|
|
|
|
input_price: 0.27
|
|
|
|
|
output_price: 0.4
|
|
|
|
|
- name: deepseek/deepseek-v3.1-terminus
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 163840
|
|
|
|
|
input_price: 0.23
|
|
|
|
|
output_price: 0.9
|
|
|
|
|
- name: deepseek/deepseek-chat-v3.1
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 163840
|
|
|
|
|
input_price: 0.2
|
|
|
|
|
output_price: 0.8
|
|
|
|
|
- name: deepseek/deepseek-r1-0528
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
input_price: 0.5
|
|
|
|
|
output_price: 2.15
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
include_reasoning: true
|
|
|
|
|
input_price: 0.25
|
|
|
|
|
output_price: 0.38
|
|
|
|
|
- name: qwen/qwen3-max
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
@@ -1982,13 +1663,7 @@ list:
|
|
|
|
|
input_price: 0.29
|
|
|
|
|
output_price: 1.15
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: x-ai/grok-4
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 256000
|
|
|
|
|
input_price: 3.0
|
|
|
|
|
output_price: 15.0
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: x-ai/grok-4-fast
|
|
|
|
|
- name: x-ai/grok-4.1-fast
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 2000000
|
|
|
|
|
input_price: 0.2
|
|
|
|
|
@@ -2044,14 +1719,6 @@ list:
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
include_reasoning: true
|
|
|
|
|
- name: perplexity/sonar-reasoning
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 127000
|
|
|
|
|
input_price: 1.0
|
|
|
|
|
output_price: 5.0
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
include_reasoning: true
|
|
|
|
|
- name: perplexity/sonar-deep-research
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
@@ -2060,17 +1727,24 @@ list:
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
include_reasoning: true
|
|
|
|
|
- name: minimax/minimax-m2
|
|
|
|
|
- name: minimax/minimax-m2.1
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 196608
|
|
|
|
|
input_price: 0.15
|
|
|
|
|
output_price: 0.45
|
|
|
|
|
- name: z-ai/glm-4.6
|
|
|
|
|
input_price: 0.12
|
|
|
|
|
output_price: 0.48
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: z-ai/glm-4.7
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 202752
|
|
|
|
|
input_price: 0.5
|
|
|
|
|
output_price: 1.75
|
|
|
|
|
input_price: 0.16
|
|
|
|
|
output_price: 0.8
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: z-ai/glm-4.6v
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.3
|
|
|
|
|
output_price: 0.9
|
|
|
|
|
supports_vision: true
|
|
|
|
|
- provider: github
|
|
|
|
|
models:
|
|
|
|
|
- name: gpt-5
|
|
|
|
|
@@ -2079,12 +1753,6 @@ list:
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-5-chat
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
max_output_tokens: 128000
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-5-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 400000
|
|
|
|
|
@@ -2103,100 +1771,11 @@ list:
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-4.1-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1047576
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-4.1-nano
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 1047576
|
|
|
|
|
max_output_tokens: 32768
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-4o
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
max_output_tokens: 16384
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: gpt-4o-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 128000
|
|
|
|
|
max_output_tokens: 16384
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: o4-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o4-mini-high
|
|
|
|
|
type: chat
|
|
|
|
|
real_name: o4-mini
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
reasoning_effort: high
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o3
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o3-high
|
|
|
|
|
type: chat
|
|
|
|
|
real_name: o3
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
reasoning_effort: high
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o3-mini
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: o3-mini-high
|
|
|
|
|
type: chat
|
|
|
|
|
real_name: o3-mini
|
|
|
|
|
max_input_tokens: 200000
|
|
|
|
|
patch:
|
|
|
|
|
body:
|
|
|
|
|
reasoning_effort: high
|
|
|
|
|
max_tokens: null
|
|
|
|
|
temperature: null
|
|
|
|
|
top_p: null
|
|
|
|
|
supports_vision: true
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
system_prompt_prefix: Formatting re-enabled
|
|
|
|
|
- name: text-embedding-3-large
|
|
|
|
|
type: embedding
|
|
|
|
|
max_tokens_per_chunk: 8191
|
|
|
|
|
@@ -2338,26 +1917,12 @@ list:
|
|
|
|
|
input_price: 0.18
|
|
|
|
|
output_price: 0.69
|
|
|
|
|
supports_vision: true
|
|
|
|
|
- name: deepseek-ai/DeepSeek-V3.2-Exp
|
|
|
|
|
- name: deepseek-ai/DeepSeek-V3.2
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 163840
|
|
|
|
|
input_price: 0.27
|
|
|
|
|
output_price: 0.4
|
|
|
|
|
- name: deepseek-ai/DeepSeek-V3.1-Terminus
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 163840
|
|
|
|
|
input_price: 0.27
|
|
|
|
|
output_price: 1.0
|
|
|
|
|
- name: deepseek-ai/DeepSeek-V3.1
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 163840
|
|
|
|
|
input_price: 0.3
|
|
|
|
|
output_price: 1.0
|
|
|
|
|
- name: deepseek-ai/DeepSeek-R1-0528
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 163840
|
|
|
|
|
input_price: 0.5
|
|
|
|
|
output_price: 2.15
|
|
|
|
|
input_price: 0.26
|
|
|
|
|
output_price: 0.39
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: google/gemma-3-27b-it
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
@@ -2380,12 +1945,24 @@ list:
|
|
|
|
|
input_price: 0.55
|
|
|
|
|
output_price: 2.5
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: zai-org/GLM-4.6
|
|
|
|
|
- name: MiniMaxAI/MiniMax-M2.1
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 262144
|
|
|
|
|
input_price: 0.28
|
|
|
|
|
output_price: 1.2
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: zai-org/GLM-4.7
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 202752
|
|
|
|
|
input_price: 0.6
|
|
|
|
|
output_price: 1.9
|
|
|
|
|
input_price: 0.43
|
|
|
|
|
output_price: 1.75
|
|
|
|
|
supports_function_calling: true
|
|
|
|
|
- name: zai-org/GLM-4.6V
|
|
|
|
|
type: chat
|
|
|
|
|
max_input_tokens: 131072
|
|
|
|
|
input_price: 0.3
|
|
|
|
|
output_price: 0.9
|
|
|
|
|
supports_vision: true
|
|
|
|
|
- name: BAAI/bge-large-en-v1.5
|
|
|
|
|
type: embedding
|
|
|
|
|
input_price: 0.01
|
|
|
|
|
|