@@ -5643,7 +5643,8 @@ paths:
56435643 description : |
56445644 Filter hardware configurations by model compatibility. When provided,
56455645 the response includes availability status for each compatible configuration.
5646- example : meta-llama/Llama-3-70b-chat-hf
5646+ [See all of Together AI's dedicated models](https://docs.together.ai/docs/dedicated-models)
5647+ example : deepseek-ai/DeepSeek-R1
56475648 responses :
56485649 ' 200 ' :
56495650 description : ' List of available hardware configurations'
@@ -9097,20 +9098,11 @@ components:
90979098 items :
90989099 $ref : ' #/components/schemas/ChatCompletionMessageParam'
90999100 model :
9101+ type : string
91009102 description : >
91019103 The name of the model to query.<br>
91029104 <br>
91039105 [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
9104- example : Qwen/Qwen3.5-9B
9105- anyOf :
9106- - type : string
9107- enum :
9108- - Qwen/Qwen2.5-72B-Instruct-Turbo
9109- - Qwen/Qwen2.5-7B-Instruct-Turbo
9110- - Qwen/Qwen3.5-9B
9111- - meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
9112- - meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo
9113- - type : string
91149106 max_tokens :
91159107 type : integer
91169108 description : The maximum number of tokens to generate.
@@ -11313,18 +11305,15 @@ components:
1131311305 display_name :
1131411306 type : string
1131511307 description : A human-readable name for the endpoint
11316- examples :
11317- - My Llama3 70b endpoint
11308+ example : My Llama3 70b endpoint
1131811309 model :
1131911310 type : string
1132011311 description : The model to deploy on this endpoint
11321- examples :
11322- - meta-llama/Llama-3-8b-chat-hf
11312+ example : deepseek-ai/DeepSeek-R1
1132311313 hardware :
1132411314 type : string
1132511315 description : The hardware configuration to use for this endpoint
11326- examples :
11327- - 1x_nvidia_a100_80gb_sxm
11316+ example : 1x_nvidia_a100_80gb_sxm
1132811317 autoscaling :
1132911318 $ref : ' #/components/schemas/Autoscaling'
1133011319 description : Configuration for automatic scaling of the endpoint
@@ -11380,19 +11369,19 @@ components:
1138011369 name :
1138111370 type : string
1138211371 description : System name for the endpoint
11383- example : devuser/meta-llama/Llama-3-8b-chat-hf -a32b82a1
11372+ example : devuser/deepseek-ai/DeepSeek-R1 -a32b82a1
1138411373 display_name :
1138511374 type : string
1138611375 description : Human-readable name for the endpoint
11387- example : My Llama3 70b endpoint
11376+ example : My DeepSeek R1 endpoint
1138811377 model :
1138911378 type : string
1139011379 description : The model deployed on this endpoint
11391- example : meta-llama/Llama-3-8b-chat-hf
11380+ example : deepseek-ai/DeepSeek-R1
1139211381 hardware :
1139311382 type : string
1139411383 description : The hardware configuration used for this endpoint
11395- example : 1x_nvidia_a100_80gb_sxm
11384+ example : 8x_nvidia_h200_140gb_sxm
1139611385 type :
1139711386 type : string
1139811387 enum :
@@ -12069,7 +12058,7 @@ components:
1206912058 model :
1207012059 type : string
1207112060 description : Name of the judge model
12072- example : ' meta-llama/Llama-3-70B-Instruct-Turbo '
12061+ example : ' Qwen/Qwen3.5-9B '
1207312062 system_template :
1207412063 type : string
1207512064 description : System prompt template for the judge
@@ -12104,7 +12093,7 @@ components:
1210412093 model :
1210512094 type : string
1210612095 description : Name of the model to evaluate
12107- example : ' meta-llama/Llama-3-70B-Instruct-Turbo '
12096+ example : ' Qwen/Qwen3.5-9B '
1210812097 max_tokens :
1210912098 type : integer
1211012099 minimum : 1
0 commit comments