Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions .github/workflows/conformance.yml
Original file line number Diff line number Diff line change
Expand Up @@ -43,9 +43,9 @@ jobs:
# Check if we should skip conformance testing due to breaking changes
- name: Check if conformance test should be skipped
id: skip-check
env:
PR_TITLE: ${{ github.event.pull_request.title }}
run: |
PR_TITLE="${{ github.event.pull_request.title }}"

# Skip if title contains "!:" indicating breaking change (like "feat!:")
if [[ "$PR_TITLE" == *"!:"* ]]; then
echo "skip=true" >> $GITHUB_OUTPUT
Expand Down
23 changes: 6 additions & 17 deletions docs/static/deprecated-llama-stack-spec.html
Original file line number Diff line number Diff line change
Expand Up @@ -1527,7 +1527,7 @@
"content": {
"application/json": {
"schema": {
"$ref": "#/components/schemas/OpenAIChatCompletionRequest"
"$ref": "#/components/schemas/OpenAIChatCompletionRequestWithExtraBody"
}
}
},
Expand Down Expand Up @@ -1617,7 +1617,7 @@
"content": {
"application/json": {
"schema": {
"$ref": "#/components/schemas/OpenAICompletionRequest"
"$ref": "#/components/schemas/OpenAICompletionRequestWithExtraBody"
}
}
},
Expand Down Expand Up @@ -7522,7 +7522,7 @@
"title": "OpenAIResponseFormatText",
"description": "Text response format for OpenAI-compatible chat completion requests."
},
"OpenAIChatCompletionRequest": {
"OpenAIChatCompletionRequestWithExtraBody": {
"type": "object",
"properties": {
"model": {
Expand Down Expand Up @@ -7769,7 +7769,7 @@
"model",
"messages"
],
"title": "OpenAIChatCompletionRequest",
"title": "OpenAIChatCompletionRequestWithExtraBody",
"description": "Request parameters for OpenAI-compatible chat completion endpoint."
},
"OpenAIChatCompletion": {
Expand Down Expand Up @@ -7966,7 +7966,7 @@
],
"title": "OpenAICompletionWithInputMessages"
},
"OpenAICompletionRequest": {
"OpenAICompletionRequestWithExtraBody": {
"type": "object",
"properties": {
"model": {
Expand Down Expand Up @@ -8097,17 +8097,6 @@
"type": "string",
"description": "(Optional) The user to use."
},
"guided_choice": {
"type": "array",
"items": {
"type": "string"
},
"description": "(Optional) vLLM-specific parameter for guided generation with a list of choices."
},
"prompt_logprobs": {
"type": "integer",
"description": "(Optional) vLLM-specific parameter for number of log probabilities to return for prompt tokens."
},
"suffix": {
"type": "string",
"description": "(Optional) The suffix that should be appended to the completion."
Expand All @@ -8118,7 +8107,7 @@
"model",
"prompt"
],
"title": "OpenAICompletionRequest",
"title": "OpenAICompletionRequestWithExtraBody",
"description": "Request parameters for OpenAI-compatible completion endpoint."
},
"OpenAICompletion": {
Expand Down
24 changes: 6 additions & 18 deletions docs/static/deprecated-llama-stack-spec.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -1098,7 +1098,7 @@ paths:
content:
application/json:
schema:
$ref: '#/components/schemas/OpenAIChatCompletionRequest'
$ref: '#/components/schemas/OpenAIChatCompletionRequestWithExtraBody'
required: true
deprecated: true
/v1/openai/v1/chat/completions/{completion_id}:
Expand Down Expand Up @@ -1167,7 +1167,7 @@ paths:
content:
application/json:
schema:
$ref: '#/components/schemas/OpenAICompletionRequest'
$ref: '#/components/schemas/OpenAICompletionRequestWithExtraBody'
required: true
deprecated: true
/v1/openai/v1/embeddings:
Expand Down Expand Up @@ -5575,7 +5575,7 @@ components:
title: OpenAIResponseFormatText
description: >-
Text response format for OpenAI-compatible chat completion requests.
OpenAIChatCompletionRequest:
OpenAIChatCompletionRequestWithExtraBody:
type: object
properties:
model:
Expand Down Expand Up @@ -5717,7 +5717,7 @@ components:
required:
- model
- messages
title: OpenAIChatCompletionRequest
title: OpenAIChatCompletionRequestWithExtraBody
description: >-
Request parameters for OpenAI-compatible chat completion endpoint.
OpenAIChatCompletion:
Expand Down Expand Up @@ -5885,7 +5885,7 @@ components:
- model
- input_messages
title: OpenAICompletionWithInputMessages
OpenAICompletionRequest:
OpenAICompletionRequestWithExtraBody:
type: object
properties:
model:
Expand Down Expand Up @@ -5973,18 +5973,6 @@ components:
user:
type: string
description: (Optional) The user to use.
guided_choice:
type: array
items:
type: string
description: >-
(Optional) vLLM-specific parameter for guided generation with a list of
choices.
prompt_logprobs:
type: integer
description: >-
(Optional) vLLM-specific parameter for number of log probabilities to
return for prompt tokens.
suffix:
type: string
description: >-
Expand All @@ -5993,7 +5981,7 @@ components:
required:
- model
- prompt
title: OpenAICompletionRequest
title: OpenAICompletionRequestWithExtraBody
description: >-
Request parameters for OpenAI-compatible completion endpoint.
OpenAICompletion:
Expand Down
23 changes: 6 additions & 17 deletions docs/static/llama-stack-spec.html
Original file line number Diff line number Diff line change
Expand Up @@ -153,7 +153,7 @@
"content": {
"application/json": {
"schema": {
"$ref": "#/components/schemas/OpenAIChatCompletionRequest"
"$ref": "#/components/schemas/OpenAIChatCompletionRequestWithExtraBody"
}
}
},
Expand Down Expand Up @@ -243,7 +243,7 @@
"content": {
"application/json": {
"schema": {
"$ref": "#/components/schemas/OpenAICompletionRequest"
"$ref": "#/components/schemas/OpenAICompletionRequestWithExtraBody"
}
}
},
Expand Down Expand Up @@ -5018,7 +5018,7 @@
"title": "OpenAIResponseFormatText",
"description": "Text response format for OpenAI-compatible chat completion requests."
},
"OpenAIChatCompletionRequest": {
"OpenAIChatCompletionRequestWithExtraBody": {
"type": "object",
"properties": {
"model": {
Expand Down Expand Up @@ -5265,7 +5265,7 @@
"model",
"messages"
],
"title": "OpenAIChatCompletionRequest",
"title": "OpenAIChatCompletionRequestWithExtraBody",
"description": "Request parameters for OpenAI-compatible chat completion endpoint."
},
"OpenAIChatCompletion": {
Expand Down Expand Up @@ -5462,7 +5462,7 @@
],
"title": "OpenAICompletionWithInputMessages"
},
"OpenAICompletionRequest": {
"OpenAICompletionRequestWithExtraBody": {
"type": "object",
"properties": {
"model": {
Expand Down Expand Up @@ -5593,17 +5593,6 @@
"type": "string",
"description": "(Optional) The user to use."
},
"guided_choice": {
"type": "array",
"items": {
"type": "string"
},
"description": "(Optional) vLLM-specific parameter for guided generation with a list of choices."
},
"prompt_logprobs": {
"type": "integer",
"description": "(Optional) vLLM-specific parameter for number of log probabilities to return for prompt tokens."
},
"suffix": {
"type": "string",
"description": "(Optional) The suffix that should be appended to the completion."
Expand All @@ -5614,7 +5603,7 @@
"model",
"prompt"
],
"title": "OpenAICompletionRequest",
"title": "OpenAICompletionRequestWithExtraBody",
"description": "Request parameters for OpenAI-compatible completion endpoint."
},
"OpenAICompletion": {
Expand Down
24 changes: 6 additions & 18 deletions docs/static/llama-stack-spec.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -98,7 +98,7 @@ paths:
content:
application/json:
schema:
$ref: '#/components/schemas/OpenAIChatCompletionRequest'
$ref: '#/components/schemas/OpenAIChatCompletionRequestWithExtraBody'
required: true
deprecated: false
/v1/chat/completions/{completion_id}:
Expand Down Expand Up @@ -167,7 +167,7 @@ paths:
content:
application/json:
schema:
$ref: '#/components/schemas/OpenAICompletionRequest'
$ref: '#/components/schemas/OpenAICompletionRequestWithExtraBody'
required: true
deprecated: false
/v1/conversations:
Expand Down Expand Up @@ -3824,7 +3824,7 @@ components:
title: OpenAIResponseFormatText
description: >-
Text response format for OpenAI-compatible chat completion requests.
OpenAIChatCompletionRequest:
OpenAIChatCompletionRequestWithExtraBody:
type: object
properties:
model:
Expand Down Expand Up @@ -3966,7 +3966,7 @@ components:
required:
- model
- messages
title: OpenAIChatCompletionRequest
title: OpenAIChatCompletionRequestWithExtraBody
description: >-
Request parameters for OpenAI-compatible chat completion endpoint.
OpenAIChatCompletion:
Expand Down Expand Up @@ -4134,7 +4134,7 @@ components:
- model
- input_messages
title: OpenAICompletionWithInputMessages
OpenAICompletionRequest:
OpenAICompletionRequestWithExtraBody:
type: object
properties:
model:
Expand Down Expand Up @@ -4222,18 +4222,6 @@ components:
user:
type: string
description: (Optional) The user to use.
guided_choice:
type: array
items:
type: string
description: >-
(Optional) vLLM-specific parameter for guided generation with a list of
choices.
prompt_logprobs:
type: integer
description: >-
(Optional) vLLM-specific parameter for number of log probabilities to
return for prompt tokens.
suffix:
type: string
description: >-
Expand All @@ -4242,7 +4230,7 @@ components:
required:
- model
- prompt
title: OpenAICompletionRequest
title: OpenAICompletionRequestWithExtraBody
description: >-
Request parameters for OpenAI-compatible completion endpoint.
OpenAICompletion:
Expand Down
23 changes: 6 additions & 17 deletions docs/static/stainless-llama-stack-spec.html
Original file line number Diff line number Diff line change
Expand Up @@ -153,7 +153,7 @@
"content": {
"application/json": {
"schema": {
"$ref": "#/components/schemas/OpenAIChatCompletionRequest"
"$ref": "#/components/schemas/OpenAIChatCompletionRequestWithExtraBody"
}
}
},
Expand Down Expand Up @@ -243,7 +243,7 @@
"content": {
"application/json": {
"schema": {
"$ref": "#/components/schemas/OpenAICompletionRequest"
"$ref": "#/components/schemas/OpenAICompletionRequestWithExtraBody"
}
}
},
Expand Down Expand Up @@ -7027,7 +7027,7 @@
"title": "OpenAIResponseFormatText",
"description": "Text response format for OpenAI-compatible chat completion requests."
},
"OpenAIChatCompletionRequest": {
"OpenAIChatCompletionRequestWithExtraBody": {
"type": "object",
"properties": {
"model": {
Expand Down Expand Up @@ -7274,7 +7274,7 @@
"model",
"messages"
],
"title": "OpenAIChatCompletionRequest",
"title": "OpenAIChatCompletionRequestWithExtraBody",
"description": "Request parameters for OpenAI-compatible chat completion endpoint."
},
"OpenAIChatCompletion": {
Expand Down Expand Up @@ -7471,7 +7471,7 @@
],
"title": "OpenAICompletionWithInputMessages"
},
"OpenAICompletionRequest": {
"OpenAICompletionRequestWithExtraBody": {
"type": "object",
"properties": {
"model": {
Expand Down Expand Up @@ -7602,17 +7602,6 @@
"type": "string",
"description": "(Optional) The user to use."
},
"guided_choice": {
"type": "array",
"items": {
"type": "string"
},
"description": "(Optional) vLLM-specific parameter for guided generation with a list of choices."
},
"prompt_logprobs": {
"type": "integer",
"description": "(Optional) vLLM-specific parameter for number of log probabilities to return for prompt tokens."
},
"suffix": {
"type": "string",
"description": "(Optional) The suffix that should be appended to the completion."
Expand All @@ -7623,7 +7612,7 @@
"model",
"prompt"
],
"title": "OpenAICompletionRequest",
"title": "OpenAICompletionRequestWithExtraBody",
"description": "Request parameters for OpenAI-compatible completion endpoint."
},
"OpenAICompletion": {
Expand Down
Loading
Loading