diff --git a/.stats.yml b/.stats.yml index b2aff63..2022b62 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,2 +1,2 @@ configured_endpoints: 16 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/togetherai%2FTogetherAI-1677223d9bf95220f0e49c6c62130fc30a936ffc69f2a507a68b714ab6099e8f.yml +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/togetherai%2FTogetherAI-5e0a37829859a7e57d29be0ae29e5a1094a9f5ab1ca47c8ed224b77f74a3ccad.yml diff --git a/src/resources/chat/completions.ts b/src/resources/chat/completions.ts index 6fa30f0..9e38908 100644 --- a/src/resources/chat/completions.ts +++ b/src/resources/chat/completions.ts @@ -250,18 +250,14 @@ export interface CompletionCreateParamsBase { /** * The name of the model to query. - * - * Popular models: - * - * - meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo - * - meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo - * - meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo - * - Qwen/Qwen2.5-7B-Instruct-Turbo - * - Qwen/Qwen2.5-72B-Instruct-Turbo - * - * [See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models) */ - model: string; + model: + | 'Qwen/Qwen2.5-72B-Instruct-Turbo' + | 'Qwen/Qwen2.5-7B-Instruct-Turbo' + | 'meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo' + | 'meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo' + | 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo' + | (string & {}); /** * If true, the response will contain the prompt. Can be used with `logprobs` to diff --git a/tests/api-resources/chat/completions.test.ts b/tests/api-resources/chat/completions.test.ts index 909e6d4..4769c2c 100644 --- a/tests/api-resources/chat/completions.test.ts +++ b/tests/api-resources/chat/completions.test.ts @@ -16,7 +16,7 @@ describe('resource completions', () => { { content: 'content', role: 'system' }, { content: 'content', role: 'system' }, ], - model: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo', + model: 'Qwen/Qwen2.5-72B-Instruct-Turbo', }); const rawResponse = await responsePromise.asResponse(); expect(rawResponse).toBeInstanceOf(Response); @@ -34,7 +34,7 @@ describe('resource completions', () => { { content: 'content', role: 'system' }, { content: 'content', role: 'system' }, ], - model: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo', + model: 'Qwen/Qwen2.5-72B-Instruct-Turbo', echo: true, frequency_penalty: 0, function_call: 'none',