Skip to content

Commit 08c075d

Browse files
authored
chore: Remove meta-llama from serverless examples due to deprecation (#228)
1 parent 152ef85 commit 08c075d

1 file changed

Lines changed: 12 additions & 23 deletions

File tree

openapi.yaml

Lines changed: 12 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -5643,7 +5643,8 @@ paths:
56435643
description: |
56445644
Filter hardware configurations by model compatibility. When provided,
56455645
the response includes availability status for each compatible configuration.
5646-
example: meta-llama/Llama-3-70b-chat-hf
5646+
[See all of Together AI's dedicated models](https://docs.together.ai/docs/dedicated-models)
5647+
example: deepseek-ai/DeepSeek-R1
56475648
responses:
56485649
'200':
56495650
description: 'List of available hardware configurations'
@@ -9097,20 +9098,11 @@ components:
90979098
items:
90989099
$ref: '#/components/schemas/ChatCompletionMessageParam'
90999100
model:
9101+
type: string
91009102
description: >
91019103
The name of the model to query.<br>
91029104
<br>
91039105
[See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
9104-
example: Qwen/Qwen3.5-9B
9105-
anyOf:
9106-
- type: string
9107-
enum:
9108-
- Qwen/Qwen2.5-72B-Instruct-Turbo
9109-
- Qwen/Qwen2.5-7B-Instruct-Turbo
9110-
- Qwen/Qwen3.5-9B
9111-
- meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
9112-
- meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo
9113-
- type: string
91149106
max_tokens:
91159107
type: integer
91169108
description: The maximum number of tokens to generate.
@@ -11313,18 +11305,15 @@ components:
1131311305
display_name:
1131411306
type: string
1131511307
description: A human-readable name for the endpoint
11316-
examples:
11317-
- My Llama3 70b endpoint
11308+
example: My Llama3 70b endpoint
1131811309
model:
1131911310
type: string
1132011311
description: The model to deploy on this endpoint
11321-
examples:
11322-
- meta-llama/Llama-3-8b-chat-hf
11312+
example: deepseek-ai/DeepSeek-R1
1132311313
hardware:
1132411314
type: string
1132511315
description: The hardware configuration to use for this endpoint
11326-
examples:
11327-
- 1x_nvidia_a100_80gb_sxm
11316+
example: 1x_nvidia_a100_80gb_sxm
1132811317
autoscaling:
1132911318
$ref: '#/components/schemas/Autoscaling'
1133011319
description: Configuration for automatic scaling of the endpoint
@@ -11380,19 +11369,19 @@ components:
1138011369
name:
1138111370
type: string
1138211371
description: System name for the endpoint
11383-
example: devuser/meta-llama/Llama-3-8b-chat-hf-a32b82a1
11372+
example: devuser/deepseek-ai/DeepSeek-R1-a32b82a1
1138411373
display_name:
1138511374
type: string
1138611375
description: Human-readable name for the endpoint
11387-
example: My Llama3 70b endpoint
11376+
example: My DeepSeek R1 endpoint
1138811377
model:
1138911378
type: string
1139011379
description: The model deployed on this endpoint
11391-
example: meta-llama/Llama-3-8b-chat-hf
11380+
example: deepseek-ai/DeepSeek-R1
1139211381
hardware:
1139311382
type: string
1139411383
description: The hardware configuration used for this endpoint
11395-
example: 1x_nvidia_a100_80gb_sxm
11384+
example: 8x_nvidia_h200_140gb_sxm
1139611385
type:
1139711386
type: string
1139811387
enum:
@@ -12069,7 +12058,7 @@ components:
1206912058
model:
1207012059
type: string
1207112060
description: Name of the judge model
12072-
example: 'meta-llama/Llama-3-70B-Instruct-Turbo'
12061+
example: 'Qwen/Qwen3.5-9B'
1207312062
system_template:
1207412063
type: string
1207512064
description: System prompt template for the judge
@@ -12104,7 +12093,7 @@ components:
1210412093
model:
1210512094
type: string
1210612095
description: Name of the model to evaluate
12107-
example: 'meta-llama/Llama-3-70B-Instruct-Turbo'
12096+
example: 'Qwen/Qwen3.5-9B'
1210812097
max_tokens:
1210912098
type: integer
1211012099
minimum: 1

0 commit comments

Comments
 (0)