From 28e3839c5d6bbd2b2cd048e1b7b297d3a7c52eda Mon Sep 17 00:00:00 2001 From: Akhil Madhu Menon Date: Wed, 8 Apr 2026 18:04:59 +0530 Subject: [PATCH] Updated GET /models endpoint --- openapi.yaml | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/openapi.yaml b/openapi.yaml index 04296fa5..0e7441ae 100644 --- a/openapi.yaml +++ b/openapi.yaml @@ -5012,31 +5012,31 @@ paths: - in: query name: ai_service required: false - description: Filter models by the AI service (e.g., 'openai', 'anthropic'). + description: Filter models by AI service (e.g., 'openai', 'anthropic'). schema: type: string - in: query name: provider required: false - description: Filter models by the provider. + description: Filter models by virtual key slug (the provider segment in model id `@provider/model`). schema: type: string - in: query name: limit required: false - description: The maximum number of models to return. + description: Maximum number of models to return per page. schema: type: integer - in: query name: offset required: false - description: The number of models to skip before starting to collect the result set. + description: Number of models to skip before collecting the result set. schema: type: integer - in: query name: sort required: false - description: The field to sort the results by. + description: The field to sort the results by (`provider` sorts by virtual key slug). schema: type: string enum: [name, provider, ai_service] @@ -5071,13 +5071,13 @@ paths: label: Default source: | # Example of sending a query parameter in the URL - curl 'https://api.portkey.ai/v1/models?provider=openai' \ + curl 'https://api.portkey.ai/v1/models?ai_service=openai' \ -H "x-portkey-api-key: $PORTKEY_API_KEY" - lang: curl label: Self-Hosted source: | # Example of sending a query parameter in the URL - curl 'https://YOUR_SELF_HOSTED_URL/models?provider=openai' \ + curl 'https://YOUR_SELF_HOSTED_URL/models?ai_service=openai' \ -H "x-portkey-api-key: $PORTKEY_API_KEY" - lang: python label: Default @@ -5090,7 +5090,7 @@ paths: # Example of sending query parameters via extra_query models = client.models.list( - extra_query={"provider": "openai"} + extra_query={"ai_service": "openai"} ) print(models) - lang: python @@ -5105,7 +5105,7 @@ paths: # Example of sending query parameters via extra_query models = client.models.list( - extra_query={"provider": "openai"} + extra_query={"ai_service": "openai"} ) print(models) - lang: javascript @@ -5120,7 +5120,7 @@ paths: async function main() { // Example of sending query parameters in the list method const list = await client.models.list({ - provider: "openai" + ai_service: "openai" }); console.log(list); } @@ -5138,7 +5138,7 @@ paths: async function main() { // Example of sending query parameters in the list method const list = await client.models.list({ - provider: "openai" + ai_service: "openai" }); console.log(list); }