Skip to content

Commit 2fa1a33

Browse files
feat(api): manual updates
delete /v1/openai/v1/responses/{response_id} post /v1/inference/rerank get /v1/openai/v1/models delete /v1/shields/{identifier}
1 parent b1d5c7d commit 2fa1a33

File tree

17 files changed

+500
-131
lines changed

17 files changed

+500
-131
lines changed

.stats.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
configured_endpoints: 107
1+
configured_endpoints: 111
22
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-f252873ea1e1f38fd207331ef2621c511154d5be3f4076e59cc15754fc58eee4.yml
33
openapi_spec_hash: 10cbb4337a06a9fdd7d08612dd6044c3
4-
config_hash: 374d9711288576877a9fabb34e4da7b9
4+
config_hash: 0358112cc0f3d880b4d55debdbe1cfa3

api.md

Lines changed: 22 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -81,12 +81,14 @@ Types:
8181
- <code><a href="./src/resources/responses/responses.ts">ResponseObject</a></code>
8282
- <code><a href="./src/resources/responses/responses.ts">ResponseObjectStream</a></code>
8383
- <code><a href="./src/resources/responses/responses.ts">ResponseListResponse</a></code>
84+
- <code><a href="./src/resources/responses/responses.ts">ResponseDeleteResponse</a></code>
8485

8586
Methods:
8687

8788
- <code title="post /v1/openai/v1/responses">client.responses.<a href="./src/resources/responses/responses.ts">create</a>({ ...params }) -> ResponseObject</code>
8889
- <code title="get /v1/openai/v1/responses/{response_id}">client.responses.<a href="./src/resources/responses/responses.ts">retrieve</a>(responseId) -> ResponseObject</code>
8990
- <code title="get /v1/openai/v1/responses">client.responses.<a href="./src/resources/responses/responses.ts">list</a>({ ...params }) -> ResponseListResponsesOpenAICursorPage</code>
91+
- <code title="delete /v1/openai/v1/responses/{response_id}">client.responses.<a href="./src/resources/responses/responses.ts">delete</a>(responseId) -> ResponseDeleteResponse</code>
9092

9193
## InputItems
9294

@@ -224,6 +226,7 @@ Types:
224226
- <code><a href="./src/resources/inference.ts">EmbeddingsResponse</a></code>
225227
- <code><a href="./src/resources/inference.ts">TokenLogProbs</a></code>
226228
- <code><a href="./src/resources/inference.ts">InferenceBatchChatCompletionResponse</a></code>
229+
- <code><a href="./src/resources/inference.ts">InferenceRerankResponse</a></code>
227230

228231
Methods:
229232

@@ -232,6 +235,7 @@ Methods:
232235
- <code title="post /v1/inference/chat-completion">client.inference.<a href="./src/resources/inference.ts">chatCompletion</a>({ ...params }) -> ChatCompletionResponse</code>
233236
- <code title="post /v1/inference/completion">client.inference.<a href="./src/resources/inference.ts">completion</a>({ ...params }) -> CompletionResponse</code>
234237
- <code title="post /v1/inference/embeddings">client.inference.<a href="./src/resources/inference.ts">embeddings</a>({ ...params }) -> EmbeddingsResponse</code>
238+
- <code title="post /v1/inference/rerank">client.inference.<a href="./src/resources/inference.ts">rerank</a>({ ...params }) -> InferenceRerankResponse</code>
235239

236240
# Embeddings
237241

@@ -339,16 +343,26 @@ Methods:
339343

340344
Types:
341345

342-
- <code><a href="./src/resources/models.ts">ListModelsResponse</a></code>
343-
- <code><a href="./src/resources/models.ts">Model</a></code>
344-
- <code><a href="./src/resources/models.ts">ModelListResponse</a></code>
346+
- <code><a href="./src/resources/models/models.ts">ListModelsResponse</a></code>
347+
- <code><a href="./src/resources/models/models.ts">Model</a></code>
348+
- <code><a href="./src/resources/models/models.ts">ModelListResponse</a></code>
345349

346350
Methods:
347351

348-
- <code title="get /v1/models/{model_id}">client.models.<a href="./src/resources/models.ts">retrieve</a>(modelId) -> Model</code>
349-
- <code title="get /v1/models">client.models.<a href="./src/resources/models.ts">list</a>() -> ModelListResponse</code>
350-
- <code title="post /v1/models">client.models.<a href="./src/resources/models.ts">register</a>({ ...params }) -> Model</code>
351-
- <code title="delete /v1/models/{model_id}">client.models.<a href="./src/resources/models.ts">unregister</a>(modelId) -> void</code>
352+
- <code title="get /v1/models/{model_id}">client.models.<a href="./src/resources/models/models.ts">retrieve</a>(modelId) -> Model</code>
353+
- <code title="get /v1/models">client.models.<a href="./src/resources/models/models.ts">list</a>() -> ModelListResponse</code>
354+
- <code title="post /v1/models">client.models.<a href="./src/resources/models/models.ts">register</a>({ ...params }) -> Model</code>
355+
- <code title="delete /v1/models/{model_id}">client.models.<a href="./src/resources/models/models.ts">unregister</a>(modelId) -> void</code>
356+
357+
## OpenAI
358+
359+
Types:
360+
361+
- <code><a href="./src/resources/models/openai.ts">OpenAIListResponse</a></code>
362+
363+
Methods:
364+
365+
- <code title="get /v1/openai/v1/models">client.models.openai.<a href="./src/resources/models/openai.ts">list</a>() -> OpenAIListResponse</code>
352366

353367
# PostTraining
354368

@@ -433,6 +447,7 @@ Methods:
433447

434448
- <code title="get /v1/shields/{identifier}">client.shields.<a href="./src/resources/shields.ts">retrieve</a>(identifier) -> Shield</code>
435449
- <code title="get /v1/shields">client.shields.<a href="./src/resources/shields.ts">list</a>() -> ShieldListResponse</code>
450+
- <code title="delete /v1/shields/{identifier}">client.shields.<a href="./src/resources/shields.ts">delete</a>(identifier) -> void</code>
436451
- <code title="post /v1/shields">client.shields.<a href="./src/resources/shields.ts">register</a>({ ...params }) -> Shield</code>
437452

438453
# SyntheticDataGeneration

src/index.ts

Lines changed: 13 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -64,16 +64,11 @@ import {
6464
InferenceCompletionParamsNonStreaming,
6565
InferenceCompletionParamsStreaming,
6666
InferenceEmbeddingsParams,
67+
InferenceRerankParams,
68+
InferenceRerankResponse,
6769
TokenLogProbs,
6870
} from './resources/inference';
6971
import { HealthInfo, Inspect, ProviderInfo, RouteInfo, VersionInfo } from './resources/inspect';
70-
import {
71-
ListModelsResponse,
72-
Model,
73-
ModelListResponse,
74-
ModelRegisterParams,
75-
Models,
76-
} from './resources/models';
7772
import { CreateResponse, ModerationCreateParams, Moderations } from './resources/moderations';
7873
import { ListProvidersResponse, ProviderListResponse, Providers } from './resources/providers';
7974
import { ListRoutesResponse, RouteListResponse, Routes } from './resources/routes';
@@ -171,6 +166,13 @@ import {
171166
EvaluateResponse,
172167
Job,
173168
} from './resources/eval/eval';
169+
import {
170+
ListModelsResponse,
171+
Model,
172+
ModelListResponse,
173+
ModelRegisterParams,
174+
Models,
175+
} from './resources/models/models';
174176
import {
175177
AlgorithmConfig,
176178
ListPostTrainingJobsResponse,
@@ -183,6 +185,7 @@ import {
183185
ResponseCreateParams,
184186
ResponseCreateParamsNonStreaming,
185187
ResponseCreateParamsStreaming,
188+
ResponseDeleteResponse,
186189
ResponseListParams,
187190
ResponseListResponse,
188191
ResponseListResponsesOpenAICursorPage,
@@ -477,6 +480,7 @@ export declare namespace LlamaStackClient {
477480
type ResponseObject as ResponseObject,
478481
type ResponseObjectStream as ResponseObjectStream,
479482
type ResponseListResponse as ResponseListResponse,
483+
type ResponseDeleteResponse as ResponseDeleteResponse,
480484
ResponseListResponsesOpenAICursorPage as ResponseListResponsesOpenAICursorPage,
481485
type ResponseCreateParams as ResponseCreateParams,
482486
type ResponseCreateParamsNonStreaming as ResponseCreateParamsNonStreaming,
@@ -537,6 +541,7 @@ export declare namespace LlamaStackClient {
537541
type EmbeddingsResponse as EmbeddingsResponse,
538542
type TokenLogProbs as TokenLogProbs,
539543
type InferenceBatchChatCompletionResponse as InferenceBatchChatCompletionResponse,
544+
type InferenceRerankResponse as InferenceRerankResponse,
540545
type InferenceBatchChatCompletionParams as InferenceBatchChatCompletionParams,
541546
type InferenceBatchCompletionParams as InferenceBatchCompletionParams,
542547
type InferenceChatCompletionParams as InferenceChatCompletionParams,
@@ -546,6 +551,7 @@ export declare namespace LlamaStackClient {
546551
type InferenceCompletionParamsNonStreaming as InferenceCompletionParamsNonStreaming,
547552
type InferenceCompletionParamsStreaming as InferenceCompletionParamsStreaming,
548553
type InferenceEmbeddingsParams as InferenceEmbeddingsParams,
554+
type InferenceRerankParams as InferenceRerankParams,
549555
};
550556

551557
export {

src/resources/index.ts

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,6 +69,7 @@ export {
6969
type EmbeddingsResponse,
7070
type TokenLogProbs,
7171
type InferenceBatchChatCompletionResponse,
72+
type InferenceRerankResponse,
7273
type InferenceBatchChatCompletionParams,
7374
type InferenceBatchCompletionParams,
7475
type InferenceChatCompletionParams,
@@ -78,6 +79,7 @@ export {
7879
type InferenceCompletionParamsNonStreaming,
7980
type InferenceCompletionParamsStreaming,
8081
type InferenceEmbeddingsParams,
82+
type InferenceRerankParams,
8183
} from './inference';
8284
export { Inspect, type HealthInfo, type ProviderInfo, type RouteInfo, type VersionInfo } from './inspect';
8385
export {
@@ -86,7 +88,7 @@ export {
8688
type Model,
8789
type ModelListResponse,
8890
type ModelRegisterParams,
89-
} from './models';
91+
} from './models/models';
9092
export { Moderations, type CreateResponse, type ModerationCreateParams } from './moderations';
9193
export {
9294
PostTraining,
@@ -103,6 +105,7 @@ export {
103105
type ResponseObject,
104106
type ResponseObjectStream,
105107
type ResponseListResponse,
108+
type ResponseDeleteResponse,
106109
type ResponseCreateParams,
107110
type ResponseCreateParamsNonStreaming,
108111
type ResponseCreateParamsStreaming,

src/resources/inference.ts

Lines changed: 167 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -95,6 +95,20 @@ export class Inference extends APIResource {
9595
): Core.APIPromise<EmbeddingsResponse> {
9696
return this._client.post('/v1/inference/embeddings', { body, ...options });
9797
}
98+
99+
/**
100+
* Rerank a list of documents based on their relevance to a query.
101+
*/
102+
rerank(
103+
body: InferenceRerankParams,
104+
options?: Core.RequestOptions,
105+
): Core.APIPromise<InferenceRerankResponse> {
106+
return (
107+
this._client.post('/v1/inference/rerank', { body, ...options }) as Core.APIPromise<{
108+
data: InferenceRerankResponse;
109+
}>
110+
)._thenUnwrap((obj) => obj.data);
111+
}
98112
}
99113

100114
/**
@@ -197,6 +211,29 @@ export interface InferenceBatchChatCompletionResponse {
197211
batch: Array<Shared.ChatCompletionResponse>;
198212
}
199213

214+
/**
215+
* List of rerank result objects, sorted by relevance score (descending)
216+
*/
217+
export type InferenceRerankResponse = Array<InferenceRerankResponse.InferenceRerankResponseItem>;
218+
219+
export namespace InferenceRerankResponse {
220+
/**
221+
* A single rerank result from a reranking response.
222+
*/
223+
export interface InferenceRerankResponseItem {
224+
/**
225+
* The original index of the document in the input list
226+
*/
227+
index: number;
228+
229+
/**
230+
* The relevance score from the model output. Values are inverted when applicable
231+
* so that higher scores indicate greater relevance.
232+
*/
233+
relevance_score: number;
234+
}
235+
}
236+
200237
export interface InferenceBatchChatCompletionParams {
201238
/**
202239
* The messages to generate completions for.
@@ -575,13 +612,142 @@ export interface InferenceEmbeddingsParams {
575612
text_truncation?: 'none' | 'start' | 'end';
576613
}
577614

615+
export interface InferenceRerankParams {
616+
/**
617+
* List of items to rerank. Each item can be a string, text content part, or image
618+
* content part. Each input must not exceed the model's max input token length.
619+
*/
620+
items: Array<
621+
| string
622+
| InferenceRerankParams.OpenAIChatCompletionContentPartTextParam
623+
| InferenceRerankParams.OpenAIChatCompletionContentPartImageParam
624+
>;
625+
626+
/**
627+
* The identifier of the reranking model to use.
628+
*/
629+
model: string;
630+
631+
/**
632+
* The search query to rank items against. Can be a string, text content part, or
633+
* image content part. The input must not exceed the model's max input token
634+
* length.
635+
*/
636+
query:
637+
| string
638+
| InferenceRerankParams.OpenAIChatCompletionContentPartTextParam
639+
| InferenceRerankParams.OpenAIChatCompletionContentPartImageParam;
640+
641+
/**
642+
* (Optional) Maximum number of results to return. Default: returns all.
643+
*/
644+
max_num_results?: number;
645+
}
646+
647+
export namespace InferenceRerankParams {
648+
/**
649+
* Text content part for OpenAI-compatible chat completion messages.
650+
*/
651+
export interface OpenAIChatCompletionContentPartTextParam {
652+
/**
653+
* The text content of the message
654+
*/
655+
text: string;
656+
657+
/**
658+
* Must be "text" to identify this as text content
659+
*/
660+
type: 'text';
661+
}
662+
663+
/**
664+
* Image content part for OpenAI-compatible chat completion messages.
665+
*/
666+
export interface OpenAIChatCompletionContentPartImageParam {
667+
/**
668+
* Image URL specification and processing details
669+
*/
670+
image_url: OpenAIChatCompletionContentPartImageParam.ImageURL;
671+
672+
/**
673+
* Must be "image_url" to identify this as image content
674+
*/
675+
type: 'image_url';
676+
}
677+
678+
export namespace OpenAIChatCompletionContentPartImageParam {
679+
/**
680+
* Image URL specification and processing details
681+
*/
682+
export interface ImageURL {
683+
/**
684+
* URL of the image to include in the message
685+
*/
686+
url: string;
687+
688+
/**
689+
* (Optional) Level of detail for image processing. Can be "low", "high", or "auto"
690+
*/
691+
detail?: string;
692+
}
693+
}
694+
695+
/**
696+
* Text content part for OpenAI-compatible chat completion messages.
697+
*/
698+
export interface OpenAIChatCompletionContentPartTextParam {
699+
/**
700+
* The text content of the message
701+
*/
702+
text: string;
703+
704+
/**
705+
* Must be "text" to identify this as text content
706+
*/
707+
type: 'text';
708+
}
709+
710+
/**
711+
* Image content part for OpenAI-compatible chat completion messages.
712+
*/
713+
export interface OpenAIChatCompletionContentPartImageParam {
714+
/**
715+
* Image URL specification and processing details
716+
*/
717+
image_url: OpenAIChatCompletionContentPartImageParam.ImageURL;
718+
719+
/**
720+
* Must be "image_url" to identify this as image content
721+
*/
722+
type: 'image_url';
723+
}
724+
725+
export namespace OpenAIChatCompletionContentPartImageParam {
726+
/**
727+
* Image URL specification and processing details
728+
*/
729+
export interface ImageURL {
730+
/**
731+
* URL of the image to include in the message
732+
*/
733+
url: string;
734+
735+
/**
736+
* (Optional) Level of detail for image processing. Can be "low", "high", or "auto"
737+
*/
738+
detail?: string;
739+
}
740+
}
741+
}
742+
578743
export declare namespace Inference {
579744
export {
580745
type ChatCompletionResponseStreamChunk as ChatCompletionResponseStreamChunk,
581746
type CompletionResponse as CompletionResponse,
582747
type EmbeddingsResponse as EmbeddingsResponse,
583748
type TokenLogProbs as TokenLogProbs,
584749
type InferenceBatchChatCompletionResponse as InferenceBatchChatCompletionResponse,
750+
type InferenceRerankResponse as InferenceRerankResponse,
585751
type InferenceBatchChatCompletionParams as InferenceBatchChatCompletionParams,
586752
type InferenceBatchCompletionParams as InferenceBatchCompletionParams,
587753
type InferenceChatCompletionParams as InferenceChatCompletionParams,
@@ -591,5 +757,6 @@ export declare namespace Inference {
591757
type InferenceCompletionParamsNonStreaming as InferenceCompletionParamsNonStreaming,
592758
type InferenceCompletionParamsStreaming as InferenceCompletionParamsStreaming,
593759
type InferenceEmbeddingsParams as InferenceEmbeddingsParams,
760+
type InferenceRerankParams as InferenceRerankParams,
594761
};
595762
}

0 commit comments

Comments
 (0)