node-ejs-renderer/node_modules/openai/resources/beta/threads/runs/runs.d.ts
2024-06-09 13:55:01 -04:00

1194 lines
53 KiB
TypeScript

import * as Core from "../../../../core.js";
import { APIPromise } from "../../../../core.js";
import { APIResource } from "../../../../resource.js";
import { AssistantStream, RunCreateParamsBaseStream } from "../../../../lib/AssistantStream.js";
import { RunSubmitToolOutputsParamsStream } from "../../../../lib/AssistantStream.js";
import * as RunsAPI from "./runs.js";
import * as AssistantsAPI from "../../assistants.js";
import * as MessagesAPI from "../messages.js";
import * as ThreadsAPI from "../threads.js";
import * as StepsAPI from "./steps.js";
import { CursorPage, type CursorPageParams } from "../../../../pagination.js";
import { Stream } from "../../../../streaming.js";
export declare class Runs extends APIResource {
steps: StepsAPI.Steps;
/**
* Create a run.
*/
create(threadId: string, body: RunCreateParamsNonStreaming, options?: Core.RequestOptions): APIPromise<Run>;
create(threadId: string, body: RunCreateParamsStreaming, options?: Core.RequestOptions): APIPromise<Stream<AssistantsAPI.AssistantStreamEvent>>;
create(threadId: string, body: RunCreateParamsBase, options?: Core.RequestOptions): APIPromise<Stream<AssistantsAPI.AssistantStreamEvent> | Run>;
/**
* Retrieves a run.
*/
retrieve(threadId: string, runId: string, options?: Core.RequestOptions): Core.APIPromise<Run>;
/**
* Modifies a run.
*/
update(threadId: string, runId: string, body: RunUpdateParams, options?: Core.RequestOptions): Core.APIPromise<Run>;
/**
* Returns a list of runs belonging to a thread.
*/
list(threadId: string, query?: RunListParams, options?: Core.RequestOptions): Core.PagePromise<RunsPage, Run>;
list(threadId: string, options?: Core.RequestOptions): Core.PagePromise<RunsPage, Run>;
/**
* Cancels a run that is `in_progress`.
*/
cancel(threadId: string, runId: string, options?: Core.RequestOptions): Core.APIPromise<Run>;
/**
* A helper to create a run an poll for a terminal state. More information on Run
* lifecycles can be found here:
* https://platform.openai.com/docs/assistants/how-it-works/runs-and-run-steps
*/
createAndPoll(threadId: string, body: RunCreateParamsNonStreaming, options?: Core.RequestOptions & {
pollIntervalMs?: number;
}): Promise<Run>;
/**
* Create a Run stream
*
* @deprecated use `stream` instead
*/
createAndStream(threadId: string, body: RunCreateParamsBaseStream, options?: Core.RequestOptions): AssistantStream;
/**
* A helper to poll a run status until it reaches a terminal state. More
* information on Run lifecycles can be found here:
* https://platform.openai.com/docs/assistants/how-it-works/runs-and-run-steps
*/
poll(threadId: string, runId: string, options?: Core.RequestOptions & {
pollIntervalMs?: number;
}): Promise<Run>;
/**
* Create a Run stream
*/
stream(threadId: string, body: RunCreateParamsBaseStream, options?: Core.RequestOptions): AssistantStream;
/**
* When a run has the `status: "requires_action"` and `required_action.type` is
* `submit_tool_outputs`, this endpoint can be used to submit the outputs from the
* tool calls once they're all completed. All outputs must be submitted in a single
* request.
*/
submitToolOutputs(threadId: string, runId: string, body: RunSubmitToolOutputsParamsNonStreaming, options?: Core.RequestOptions): APIPromise<Run>;
submitToolOutputs(threadId: string, runId: string, body: RunSubmitToolOutputsParamsStreaming, options?: Core.RequestOptions): APIPromise<Stream<AssistantsAPI.AssistantStreamEvent>>;
submitToolOutputs(threadId: string, runId: string, body: RunSubmitToolOutputsParamsBase, options?: Core.RequestOptions): APIPromise<Stream<AssistantsAPI.AssistantStreamEvent> | Run>;
/**
* A helper to submit a tool output to a run and poll for a terminal run state.
* More information on Run lifecycles can be found here:
* https://platform.openai.com/docs/assistants/how-it-works/runs-and-run-steps
*/
submitToolOutputsAndPoll(threadId: string, runId: string, body: RunSubmitToolOutputsParamsNonStreaming, options?: Core.RequestOptions & {
pollIntervalMs?: number;
}): Promise<Run>;
/**
* Submit the tool outputs from a previous run and stream the run to a terminal
* state. More information on Run lifecycles can be found here:
* https://platform.openai.com/docs/assistants/how-it-works/runs-and-run-steps
*/
submitToolOutputsStream(threadId: string, runId: string, body: RunSubmitToolOutputsParamsStream, options?: Core.RequestOptions): AssistantStream;
}
export declare class RunsPage extends CursorPage<Run> {
}
/**
* Tool call objects
*/
export interface RequiredActionFunctionToolCall {
/**
* The ID of the tool call. This ID must be referenced when you submit the tool
* outputs in using the
* [Submit tool outputs to run](https://platform.openai.com/docs/api-reference/runs/submitToolOutputs)
* endpoint.
*/
id: string;
/**
* The function definition.
*/
function: RequiredActionFunctionToolCall.Function;
/**
* The type of tool call the output is required for. For now, this is always
* `function`.
*/
type: 'function';
}
export declare namespace RequiredActionFunctionToolCall {
/**
* The function definition.
*/
interface Function {
/**
* The arguments that the model expects you to pass to the function.
*/
arguments: string;
/**
* The name of the function.
*/
name: string;
}
}
/**
* Represents an execution run on a
* [thread](https://platform.openai.com/docs/api-reference/threads).
*/
export interface Run {
/**
* The identifier, which can be referenced in API endpoints.
*/
id: string;
/**
* The ID of the
* [assistant](https://platform.openai.com/docs/api-reference/assistants) used for
* execution of this run.
*/
assistant_id: string;
/**
* The Unix timestamp (in seconds) for when the run was cancelled.
*/
cancelled_at: number | null;
/**
* The Unix timestamp (in seconds) for when the run was completed.
*/
completed_at: number | null;
/**
* The Unix timestamp (in seconds) for when the run was created.
*/
created_at: number;
/**
* The Unix timestamp (in seconds) for when the run will expire.
*/
expires_at: number | null;
/**
* The Unix timestamp (in seconds) for when the run failed.
*/
failed_at: number | null;
/**
* Details on why the run is incomplete. Will be `null` if the run is not
* incomplete.
*/
incomplete_details: Run.IncompleteDetails | null;
/**
* The instructions that the
* [assistant](https://platform.openai.com/docs/api-reference/assistants) used for
* this run.
*/
instructions: string;
/**
* The last error associated with this run. Will be `null` if there are no errors.
*/
last_error: Run.LastError | null;
/**
* The maximum number of completion tokens specified to have been used over the
* course of the run.
*/
max_completion_tokens: number | null;
/**
* The maximum number of prompt tokens specified to have been used over the course
* of the run.
*/
max_prompt_tokens: number | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata: unknown | null;
/**
* The model that the
* [assistant](https://platform.openai.com/docs/api-reference/assistants) used for
* this run.
*/
model: string;
/**
* The object type, which is always `thread.run`.
*/
object: 'thread.run';
/**
* Details on the action required to continue the run. Will be `null` if no action
* is required.
*/
required_action: Run.RequiredAction | null;
/**
* Specifies the format that the model must output. Compatible with
* [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
* [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
* and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
*
* Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
* message the model generates is valid JSON.
*
* **Important:** when using JSON mode, you **must** also instruct the model to
* produce JSON yourself via a system or user message. Without this, the model may
* generate an unending stream of whitespace until the generation reaches the token
* limit, resulting in a long-running and seemingly "stuck" request. Also note that
* the message content may be partially cut off if `finish_reason="length"`, which
* indicates the generation exceeded `max_tokens` or the conversation exceeded the
* max context length.
*/
response_format: ThreadsAPI.AssistantResponseFormatOption | null;
/**
* The Unix timestamp (in seconds) for when the run was started.
*/
started_at: number | null;
/**
* The status of the run, which can be either `queued`, `in_progress`,
* `requires_action`, `cancelling`, `cancelled`, `failed`, `completed`,
* `incomplete`, or `expired`.
*/
status: RunStatus;
/**
* The ID of the [thread](https://platform.openai.com/docs/api-reference/threads)
* that was executed on as a part of this run.
*/
thread_id: string;
/**
* Controls which (if any) tool is called by the model. `none` means the model will
* not call any tools and instead generates a message. `auto` is the default value
* and means the model can pick between generating a message or calling one or more
* tools. `required` means the model must call one or more tools before responding
* to the user. Specifying a particular tool like `{"type": "file_search"}` or
* `{"type": "function", "function": {"name": "my_function"}}` forces the model to
* call that tool.
*/
tool_choice: ThreadsAPI.AssistantToolChoiceOption | null;
/**
* The list of tools that the
* [assistant](https://platform.openai.com/docs/api-reference/assistants) used for
* this run.
*/
tools: Array<AssistantsAPI.AssistantTool>;
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
truncation_strategy: Run.TruncationStrategy | null;
/**
* Usage statistics related to the run. This value will be `null` if the run is not
* in a terminal state (i.e. `in_progress`, `queued`, etc.).
*/
usage: Run.Usage | null;
/**
* The sampling temperature used for this run. If not set, defaults to 1.
*/
temperature?: number | null;
/**
* The nucleus sampling value used for this run. If not set, defaults to 1.
*/
top_p?: number | null;
}
export declare namespace Run {
/**
* Details on why the run is incomplete. Will be `null` if the run is not
* incomplete.
*/
interface IncompleteDetails {
/**
* The reason why the run is incomplete. This will point to which specific token
* limit was reached over the course of the run.
*/
reason?: 'max_completion_tokens' | 'max_prompt_tokens';
}
/**
* The last error associated with this run. Will be `null` if there are no errors.
*/
interface LastError {
/**
* One of `server_error`, `rate_limit_exceeded`, or `invalid_prompt`.
*/
code: 'server_error' | 'rate_limit_exceeded' | 'invalid_prompt';
/**
* A human-readable description of the error.
*/
message: string;
}
/**
* Details on the action required to continue the run. Will be `null` if no action
* is required.
*/
interface RequiredAction {
/**
* Details on the tool outputs needed for this run to continue.
*/
submit_tool_outputs: RequiredAction.SubmitToolOutputs;
/**
* For now, this is always `submit_tool_outputs`.
*/
type: 'submit_tool_outputs';
}
namespace RequiredAction {
/**
* Details on the tool outputs needed for this run to continue.
*/
interface SubmitToolOutputs {
/**
* A list of the relevant tool calls.
*/
tool_calls: Array<RunsAPI.RequiredActionFunctionToolCall>;
}
}
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
interface TruncationStrategy {
/**
* The truncation strategy to use for the thread. The default is `auto`. If set to
* `last_messages`, the thread will be truncated to the n most recent messages in
* the thread. When set to `auto`, messages in the middle of the thread will be
* dropped to fit the context length of the model, `max_prompt_tokens`.
*/
type: 'auto' | 'last_messages';
/**
* The number of most recent messages from the thread when constructing the context
* for the run.
*/
last_messages?: number | null;
}
/**
* Usage statistics related to the run. This value will be `null` if the run is not
* in a terminal state (i.e. `in_progress`, `queued`, etc.).
*/
interface Usage {
/**
* Number of completion tokens used over the course of the run.
*/
completion_tokens: number;
/**
* Number of prompt tokens used over the course of the run.
*/
prompt_tokens: number;
/**
* Total number of tokens used (prompt + completion).
*/
total_tokens: number;
}
}
/**
* The status of the run, which can be either `queued`, `in_progress`,
* `requires_action`, `cancelling`, `cancelled`, `failed`, `completed`,
* `incomplete`, or `expired`.
*/
export type RunStatus = 'queued' | 'in_progress' | 'requires_action' | 'cancelling' | 'cancelled' | 'failed' | 'completed' | 'incomplete' | 'expired';
export type RunCreateParams = RunCreateParamsNonStreaming | RunCreateParamsStreaming;
export interface RunCreateParamsBase {
/**
* The ID of the
* [assistant](https://platform.openai.com/docs/api-reference/assistants) to use to
* execute this run.
*/
assistant_id: string;
/**
* Appends additional instructions at the end of the instructions for the run. This
* is useful for modifying the behavior on a per-run basis without overriding other
* instructions.
*/
additional_instructions?: string | null;
/**
* Adds additional messages to the thread before creating the run.
*/
additional_messages?: Array<RunCreateParams.AdditionalMessage> | null;
/**
* Overrides the
* [instructions](https://platform.openai.com/docs/api-reference/assistants/createAssistant)
* of the assistant. This is useful for modifying the behavior on a per-run basis.
*/
instructions?: string | null;
/**
* The maximum number of completion tokens that may be used over the course of the
* run. The run will make a best effort to use only the number of completion tokens
* specified, across multiple turns of the run. If the run exceeds the number of
* completion tokens specified, the run will end with status `incomplete`. See
* `incomplete_details` for more info.
*/
max_completion_tokens?: number | null;
/**
* The maximum number of prompt tokens that may be used over the course of the run.
* The run will make a best effort to use only the number of prompt tokens
* specified, across multiple turns of the run. If the run exceeds the number of
* prompt tokens specified, the run will end with status `incomplete`. See
* `incomplete_details` for more info.
*/
max_prompt_tokens?: number | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
/**
* The ID of the [Model](https://platform.openai.com/docs/api-reference/models) to
* be used to execute this run. If a value is provided here, it will override the
* model associated with the assistant. If not, the model associated with the
* assistant will be used.
*/
model?: (string & {}) | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-0125-preview' | 'gpt-4-turbo-preview' | 'gpt-4-1106-preview' | 'gpt-4-vision-preview' | 'gpt-4' | 'gpt-4-0314' | 'gpt-4-0613' | 'gpt-4-32k' | 'gpt-4-32k-0314' | 'gpt-4-32k-0613' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-16k' | 'gpt-3.5-turbo-0613' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-16k-0613' | null;
/**
* Specifies the format that the model must output. Compatible with
* [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
* [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
* and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
*
* Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
* message the model generates is valid JSON.
*
* **Important:** when using JSON mode, you **must** also instruct the model to
* produce JSON yourself via a system or user message. Without this, the model may
* generate an unending stream of whitespace until the generation reaches the token
* limit, resulting in a long-running and seemingly "stuck" request. Also note that
* the message content may be partially cut off if `finish_reason="length"`, which
* indicates the generation exceeded `max_tokens` or the conversation exceeded the
* max context length.
*/
response_format?: ThreadsAPI.AssistantResponseFormatOption | null;
/**
* If `true`, returns a stream of events that happen during the Run as server-sent
* events, terminating when the Run enters a terminal state with a `data: [DONE]`
* message.
*/
stream?: boolean | null;
/**
* What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
* make the output more random, while lower values like 0.2 will make it more
* focused and deterministic.
*/
temperature?: number | null;
/**
* Controls which (if any) tool is called by the model. `none` means the model will
* not call any tools and instead generates a message. `auto` is the default value
* and means the model can pick between generating a message or calling one or more
* tools. `required` means the model must call one or more tools before responding
* to the user. Specifying a particular tool like `{"type": "file_search"}` or
* `{"type": "function", "function": {"name": "my_function"}}` forces the model to
* call that tool.
*/
tool_choice?: ThreadsAPI.AssistantToolChoiceOption | null;
/**
* Override the tools the assistant can use for this run. This is useful for
* modifying the behavior on a per-run basis.
*/
tools?: Array<AssistantsAPI.AssistantTool> | null;
/**
* An alternative to sampling with temperature, called nucleus sampling, where the
* model considers the results of the tokens with top_p probability mass. So 0.1
* means only the tokens comprising the top 10% probability mass are considered.
*
* We generally recommend altering this or temperature but not both.
*/
top_p?: number | null;
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
truncation_strategy?: RunCreateParams.TruncationStrategy | null;
}
export declare namespace RunCreateParams {
interface AdditionalMessage {
/**
* The text contents of the message.
*/
content: string | Array<MessagesAPI.MessageContentPartParam>;
/**
* The role of the entity that is creating the message. Allowed values include:
*
* - `user`: Indicates the message is sent by an actual user and should be used in
* most cases to represent user-generated messages.
* - `assistant`: Indicates the message is generated by the assistant. Use this
* value to insert messages from the assistant into the conversation.
*/
role: 'user' | 'assistant';
/**
* A list of files attached to the message, and the tools they should be added to.
*/
attachments?: Array<AdditionalMessage.Attachment> | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
}
namespace AdditionalMessage {
interface Attachment {
/**
* The ID of the file to attach to the message.
*/
file_id?: string;
/**
* The tools to add this file to.
*/
tools?: Array<AssistantsAPI.CodeInterpreterTool | AssistantsAPI.FileSearchTool>;
}
}
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
interface TruncationStrategy {
/**
* The truncation strategy to use for the thread. The default is `auto`. If set to
* `last_messages`, the thread will be truncated to the n most recent messages in
* the thread. When set to `auto`, messages in the middle of the thread will be
* dropped to fit the context length of the model, `max_prompt_tokens`.
*/
type: 'auto' | 'last_messages';
/**
* The number of most recent messages from the thread when constructing the context
* for the run.
*/
last_messages?: number | null;
}
type RunCreateParamsNonStreaming = RunsAPI.RunCreateParamsNonStreaming;
type RunCreateParamsStreaming = RunsAPI.RunCreateParamsStreaming;
}
export interface RunCreateParamsNonStreaming extends RunCreateParamsBase {
/**
* If `true`, returns a stream of events that happen during the Run as server-sent
* events, terminating when the Run enters a terminal state with a `data: [DONE]`
* message.
*/
stream?: false | null;
}
export interface RunCreateParamsStreaming extends RunCreateParamsBase {
/**
* If `true`, returns a stream of events that happen during the Run as server-sent
* events, terminating when the Run enters a terminal state with a `data: [DONE]`
* message.
*/
stream: true;
}
export interface RunUpdateParams {
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
}
export interface RunListParams extends CursorPageParams {
/**
* A cursor for use in pagination. `before` is an object ID that defines your place
* in the list. For instance, if you make a list request and receive 100 objects,
* ending with obj_foo, your subsequent call can include before=obj_foo in order to
* fetch the previous page of the list.
*/
before?: string;
/**
* Sort order by the `created_at` timestamp of the objects. `asc` for ascending
* order and `desc` for descending order.
*/
order?: 'asc' | 'desc';
}
export interface RunCreateAndPollParams {
/**
* The ID of the
* [assistant](https://platform.openai.com/docs/api-reference/assistants) to use to
* execute this run.
*/
assistant_id: string;
/**
* Appends additional instructions at the end of the instructions for the run. This
* is useful for modifying the behavior on a per-run basis without overriding other
* instructions.
*/
additional_instructions?: string | null;
/**
* Adds additional messages to the thread before creating the run.
*/
additional_messages?: Array<RunCreateAndPollParams.AdditionalMessage> | null;
/**
* Overrides the
* [instructions](https://platform.openai.com/docs/api-reference/assistants/createAssistant)
* of the assistant. This is useful for modifying the behavior on a per-run basis.
*/
instructions?: string | null;
/**
* The maximum number of completion tokens that may be used over the course of the
* run. The run will make a best effort to use only the number of completion tokens
* specified, across multiple turns of the run. If the run exceeds the number of
* completion tokens specified, the run will end with status `incomplete`. See
* `incomplete_details` for more info.
*/
max_completion_tokens?: number | null;
/**
* The maximum number of prompt tokens that may be used over the course of the run.
* The run will make a best effort to use only the number of prompt tokens
* specified, across multiple turns of the run. If the run exceeds the number of
* prompt tokens specified, the run will end with status `incomplete`. See
* `incomplete_details` for more info.
*/
max_prompt_tokens?: number | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
/**
* The ID of the [Model](https://platform.openai.com/docs/api-reference/models) to
* be used to execute this run. If a value is provided here, it will override the
* model associated with the assistant. If not, the model associated with the
* assistant will be used.
*/
model?: (string & {}) | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-0125-preview' | 'gpt-4-turbo-preview' | 'gpt-4-1106-preview' | 'gpt-4-vision-preview' | 'gpt-4' | 'gpt-4-0314' | 'gpt-4-0613' | 'gpt-4-32k' | 'gpt-4-32k-0314' | 'gpt-4-32k-0613' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-16k' | 'gpt-3.5-turbo-0613' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-16k-0613' | null;
/**
* Specifies the format that the model must output. Compatible with
* [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
* [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
* and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
*
* Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
* message the model generates is valid JSON.
*
* **Important:** when using JSON mode, you **must** also instruct the model to
* produce JSON yourself via a system or user message. Without this, the model may
* generate an unending stream of whitespace until the generation reaches the token
* limit, resulting in a long-running and seemingly "stuck" request. Also note that
* the message content may be partially cut off if `finish_reason="length"`, which
* indicates the generation exceeded `max_tokens` or the conversation exceeded the
* max context length.
*/
response_format?: ThreadsAPI.AssistantResponseFormatOption | null;
/**
* What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
* make the output more random, while lower values like 0.2 will make it more
* focused and deterministic.
*/
temperature?: number | null;
/**
* Controls which (if any) tool is called by the model. `none` means the model will
* not call any tools and instead generates a message. `auto` is the default value
* and means the model can pick between generating a message or calling one or more
* tools. `required` means the model must call one or more tools before responding
* to the user. Specifying a particular tool like `{"type": "file_search"}` or
* `{"type": "function", "function": {"name": "my_function"}}` forces the model to
* call that tool.
*/
tool_choice?: ThreadsAPI.AssistantToolChoiceOption | null;
/**
* Override the tools the assistant can use for this run. This is useful for
* modifying the behavior on a per-run basis.
*/
tools?: Array<AssistantsAPI.AssistantTool> | null;
/**
* An alternative to sampling with temperature, called nucleus sampling, where the
* model considers the results of the tokens with top_p probability mass. So 0.1
* means only the tokens comprising the top 10% probability mass are considered.
*
* We generally recommend altering this or temperature but not both.
*/
top_p?: number | null;
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
truncation_strategy?: RunCreateAndPollParams.TruncationStrategy | null;
}
export declare namespace RunCreateAndPollParams {
interface AdditionalMessage {
/**
* The text contents of the message.
*/
content: string | Array<MessagesAPI.MessageContentPartParam>;
/**
* The role of the entity that is creating the message. Allowed values include:
*
* - `user`: Indicates the message is sent by an actual user and should be used in
* most cases to represent user-generated messages.
* - `assistant`: Indicates the message is generated by the assistant. Use this
* value to insert messages from the assistant into the conversation.
*/
role: 'user' | 'assistant';
/**
* A list of files attached to the message, and the tools they should be added to.
*/
attachments?: Array<AdditionalMessage.Attachment> | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
}
namespace AdditionalMessage {
interface Attachment {
/**
* The ID of the file to attach to the message.
*/
file_id?: string;
/**
* The tools to add this file to.
*/
tools?: Array<AssistantsAPI.CodeInterpreterTool | AssistantsAPI.FileSearchTool>;
}
}
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
interface TruncationStrategy {
/**
* The truncation strategy to use for the thread. The default is `auto`. If set to
* `last_messages`, the thread will be truncated to the n most recent messages in
* the thread. When set to `auto`, messages in the middle of the thread will be
* dropped to fit the context length of the model, `max_prompt_tokens`.
*/
type: 'auto' | 'last_messages';
/**
* The number of most recent messages from the thread when constructing the context
* for the run.
*/
last_messages?: number | null;
}
}
export interface RunCreateAndStreamParams {
/**
* The ID of the
* [assistant](https://platform.openai.com/docs/api-reference/assistants) to use to
* execute this run.
*/
assistant_id: string;
/**
* Appends additional instructions at the end of the instructions for the run. This
* is useful for modifying the behavior on a per-run basis without overriding other
* instructions.
*/
additional_instructions?: string | null;
/**
* Adds additional messages to the thread before creating the run.
*/
additional_messages?: Array<RunCreateAndStreamParams.AdditionalMessage> | null;
/**
* Overrides the
* [instructions](https://platform.openai.com/docs/api-reference/assistants/createAssistant)
* of the assistant. This is useful for modifying the behavior on a per-run basis.
*/
instructions?: string | null;
/**
* The maximum number of completion tokens that may be used over the course of the
* run. The run will make a best effort to use only the number of completion tokens
* specified, across multiple turns of the run. If the run exceeds the number of
* completion tokens specified, the run will end with status `incomplete`. See
* `incomplete_details` for more info.
*/
max_completion_tokens?: number | null;
/**
* The maximum number of prompt tokens that may be used over the course of the run.
* The run will make a best effort to use only the number of prompt tokens
* specified, across multiple turns of the run. If the run exceeds the number of
* prompt tokens specified, the run will end with status `incomplete`. See
* `incomplete_details` for more info.
*/
max_prompt_tokens?: number | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
/**
* The ID of the [Model](https://platform.openai.com/docs/api-reference/models) to
* be used to execute this run. If a value is provided here, it will override the
* model associated with the assistant. If not, the model associated with the
* assistant will be used.
*/
model?: (string & {}) | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-0125-preview' | 'gpt-4-turbo-preview' | 'gpt-4-1106-preview' | 'gpt-4-vision-preview' | 'gpt-4' | 'gpt-4-0314' | 'gpt-4-0613' | 'gpt-4-32k' | 'gpt-4-32k-0314' | 'gpt-4-32k-0613' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-16k' | 'gpt-3.5-turbo-0613' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-16k-0613' | null;
/**
* Specifies the format that the model must output. Compatible with
* [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
* [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
* and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
*
* Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
* message the model generates is valid JSON.
*
* **Important:** when using JSON mode, you **must** also instruct the model to
* produce JSON yourself via a system or user message. Without this, the model may
* generate an unending stream of whitespace until the generation reaches the token
* limit, resulting in a long-running and seemingly "stuck" request. Also note that
* the message content may be partially cut off if `finish_reason="length"`, which
* indicates the generation exceeded `max_tokens` or the conversation exceeded the
* max context length.
*/
response_format?: ThreadsAPI.AssistantResponseFormatOption | null;
/**
* What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
* make the output more random, while lower values like 0.2 will make it more
* focused and deterministic.
*/
temperature?: number | null;
/**
* Controls which (if any) tool is called by the model. `none` means the model will
* not call any tools and instead generates a message. `auto` is the default value
* and means the model can pick between generating a message or calling one or more
* tools. `required` means the model must call one or more tools before responding
* to the user. Specifying a particular tool like `{"type": "file_search"}` or
* `{"type": "function", "function": {"name": "my_function"}}` forces the model to
* call that tool.
*/
tool_choice?: ThreadsAPI.AssistantToolChoiceOption | null;
/**
* Override the tools the assistant can use for this run. This is useful for
* modifying the behavior on a per-run basis.
*/
tools?: Array<AssistantsAPI.AssistantTool> | null;
/**
* An alternative to sampling with temperature, called nucleus sampling, where the
* model considers the results of the tokens with top_p probability mass. So 0.1
* means only the tokens comprising the top 10% probability mass are considered.
*
* We generally recommend altering this or temperature but not both.
*/
top_p?: number | null;
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
truncation_strategy?: RunCreateAndStreamParams.TruncationStrategy | null;
}
export declare namespace RunCreateAndStreamParams {
interface AdditionalMessage {
/**
* The text contents of the message.
*/
content: string | Array<MessagesAPI.MessageContentPartParam>;
/**
* The role of the entity that is creating the message. Allowed values include:
*
* - `user`: Indicates the message is sent by an actual user and should be used in
* most cases to represent user-generated messages.
* - `assistant`: Indicates the message is generated by the assistant. Use this
* value to insert messages from the assistant into the conversation.
*/
role: 'user' | 'assistant';
/**
* A list of files attached to the message, and the tools they should be added to.
*/
attachments?: Array<AdditionalMessage.Attachment> | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
}
namespace AdditionalMessage {
interface Attachment {
/**
* The ID of the file to attach to the message.
*/
file_id?: string;
/**
* The tools to add this file to.
*/
tools?: Array<AssistantsAPI.CodeInterpreterTool | AssistantsAPI.FileSearchTool>;
}
}
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
interface TruncationStrategy {
/**
* The truncation strategy to use for the thread. The default is `auto`. If set to
* `last_messages`, the thread will be truncated to the n most recent messages in
* the thread. When set to `auto`, messages in the middle of the thread will be
* dropped to fit the context length of the model, `max_prompt_tokens`.
*/
type: 'auto' | 'last_messages';
/**
* The number of most recent messages from the thread when constructing the context
* for the run.
*/
last_messages?: number | null;
}
}
export interface RunStreamParams {
/**
* The ID of the
* [assistant](https://platform.openai.com/docs/api-reference/assistants) to use to
* execute this run.
*/
assistant_id: string;
/**
* Appends additional instructions at the end of the instructions for the run. This
* is useful for modifying the behavior on a per-run basis without overriding other
* instructions.
*/
additional_instructions?: string | null;
/**
* Adds additional messages to the thread before creating the run.
*/
additional_messages?: Array<RunStreamParams.AdditionalMessage> | null;
/**
* Overrides the
* [instructions](https://platform.openai.com/docs/api-reference/assistants/createAssistant)
* of the assistant. This is useful for modifying the behavior on a per-run basis.
*/
instructions?: string | null;
/**
* The maximum number of completion tokens that may be used over the course of the
* run. The run will make a best effort to use only the number of completion tokens
* specified, across multiple turns of the run. If the run exceeds the number of
* completion tokens specified, the run will end with status `incomplete`. See
* `incomplete_details` for more info.
*/
max_completion_tokens?: number | null;
/**
* The maximum number of prompt tokens that may be used over the course of the run.
* The run will make a best effort to use only the number of prompt tokens
* specified, across multiple turns of the run. If the run exceeds the number of
* prompt tokens specified, the run will end with status `incomplete`. See
* `incomplete_details` for more info.
*/
max_prompt_tokens?: number | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
/**
* The ID of the [Model](https://platform.openai.com/docs/api-reference/models) to
* be used to execute this run. If a value is provided here, it will override the
* model associated with the assistant. If not, the model associated with the
* assistant will be used.
*/
model?: (string & {}) | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-0125-preview' | 'gpt-4-turbo-preview' | 'gpt-4-1106-preview' | 'gpt-4-vision-preview' | 'gpt-4' | 'gpt-4-0314' | 'gpt-4-0613' | 'gpt-4-32k' | 'gpt-4-32k-0314' | 'gpt-4-32k-0613' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-16k' | 'gpt-3.5-turbo-0613' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-16k-0613' | null;
/**
* Specifies the format that the model must output. Compatible with
* [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
* [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
* and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
*
* Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
* message the model generates is valid JSON.
*
* **Important:** when using JSON mode, you **must** also instruct the model to
* produce JSON yourself via a system or user message. Without this, the model may
* generate an unending stream of whitespace until the generation reaches the token
* limit, resulting in a long-running and seemingly "stuck" request. Also note that
* the message content may be partially cut off if `finish_reason="length"`, which
* indicates the generation exceeded `max_tokens` or the conversation exceeded the
* max context length.
*/
response_format?: ThreadsAPI.AssistantResponseFormatOption | null;
/**
* What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
* make the output more random, while lower values like 0.2 will make it more
* focused and deterministic.
*/
temperature?: number | null;
/**
* Controls which (if any) tool is called by the model. `none` means the model will
* not call any tools and instead generates a message. `auto` is the default value
* and means the model can pick between generating a message or calling one or more
* tools. `required` means the model must call one or more tools before responding
* to the user. Specifying a particular tool like `{"type": "file_search"}` or
* `{"type": "function", "function": {"name": "my_function"}}` forces the model to
* call that tool.
*/
tool_choice?: ThreadsAPI.AssistantToolChoiceOption | null;
/**
* Override the tools the assistant can use for this run. This is useful for
* modifying the behavior on a per-run basis.
*/
tools?: Array<AssistantsAPI.AssistantTool> | null;
/**
* An alternative to sampling with temperature, called nucleus sampling, where the
* model considers the results of the tokens with top_p probability mass. So 0.1
* means only the tokens comprising the top 10% probability mass are considered.
*
* We generally recommend altering this or temperature but not both.
*/
top_p?: number | null;
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
truncation_strategy?: RunStreamParams.TruncationStrategy | null;
}
export declare namespace RunStreamParams {
interface AdditionalMessage {
/**
* The text contents of the message.
*/
content: string | Array<MessagesAPI.MessageContentPartParam>;
/**
* The role of the entity that is creating the message. Allowed values include:
*
* - `user`: Indicates the message is sent by an actual user and should be used in
* most cases to represent user-generated messages.
* - `assistant`: Indicates the message is generated by the assistant. Use this
* value to insert messages from the assistant into the conversation.
*/
role: 'user' | 'assistant';
/**
* A list of files attached to the message, and the tools they should be added to.
*/
attachments?: Array<AdditionalMessage.Attachment> | null;
/**
* Set of 16 key-value pairs that can be attached to an object. This can be useful
* for storing additional information about the object in a structured format. Keys
* can be a maximum of 64 characters long and values can be a maxium of 512
* characters long.
*/
metadata?: unknown | null;
}
namespace AdditionalMessage {
interface Attachment {
/**
* The ID of the file to attach to the message.
*/
file_id?: string;
/**
* The tools to add this file to.
*/
tools?: Array<AssistantsAPI.CodeInterpreterTool | AssistantsAPI.FileSearchTool>;
}
}
/**
* Controls for how a thread will be truncated prior to the run. Use this to
* control the intial context window of the run.
*/
interface TruncationStrategy {
/**
* The truncation strategy to use for the thread. The default is `auto`. If set to
* `last_messages`, the thread will be truncated to the n most recent messages in
* the thread. When set to `auto`, messages in the middle of the thread will be
* dropped to fit the context length of the model, `max_prompt_tokens`.
*/
type: 'auto' | 'last_messages';
/**
* The number of most recent messages from the thread when constructing the context
* for the run.
*/
last_messages?: number | null;
}
}
export type RunSubmitToolOutputsParams = RunSubmitToolOutputsParamsNonStreaming | RunSubmitToolOutputsParamsStreaming;
export interface RunSubmitToolOutputsParamsBase {
/**
* A list of tools for which the outputs are being submitted.
*/
tool_outputs: Array<RunSubmitToolOutputsParams.ToolOutput>;
/**
* If `true`, returns a stream of events that happen during the Run as server-sent
* events, terminating when the Run enters a terminal state with a `data: [DONE]`
* message.
*/
stream?: boolean | null;
}
export declare namespace RunSubmitToolOutputsParams {
interface ToolOutput {
/**
* The output of the tool call to be submitted to continue the run.
*/
output?: string;
/**
* The ID of the tool call in the `required_action` object within the run object
* the output is being submitted for.
*/
tool_call_id?: string;
}
type RunSubmitToolOutputsParamsNonStreaming = RunsAPI.RunSubmitToolOutputsParamsNonStreaming;
type RunSubmitToolOutputsParamsStreaming = RunsAPI.RunSubmitToolOutputsParamsStreaming;
}
export interface RunSubmitToolOutputsParamsNonStreaming extends RunSubmitToolOutputsParamsBase {
/**
* If `true`, returns a stream of events that happen during the Run as server-sent
* events, terminating when the Run enters a terminal state with a `data: [DONE]`
* message.
*/
stream?: false | null;
}
export interface RunSubmitToolOutputsParamsStreaming extends RunSubmitToolOutputsParamsBase {
/**
* If `true`, returns a stream of events that happen during the Run as server-sent
* events, terminating when the Run enters a terminal state with a `data: [DONE]`
* message.
*/
stream: true;
}
export interface RunSubmitToolOutputsAndPollParams {
/**
* A list of tools for which the outputs are being submitted.
*/
tool_outputs: Array<RunSubmitToolOutputsAndPollParams.ToolOutput>;
}
export declare namespace RunSubmitToolOutputsAndPollParams {
interface ToolOutput {
/**
* The output of the tool call to be submitted to continue the run.
*/
output?: string;
/**
* The ID of the tool call in the `required_action` object within the run object
* the output is being submitted for.
*/
tool_call_id?: string;
}
}
export interface RunSubmitToolOutputsStreamParams {
/**
* A list of tools for which the outputs are being submitted.
*/
tool_outputs: Array<RunSubmitToolOutputsStreamParams.ToolOutput>;
}
export declare namespace RunSubmitToolOutputsStreamParams {
interface ToolOutput {
/**
* The output of the tool call to be submitted to continue the run.
*/
output?: string;
/**
* The ID of the tool call in the `required_action` object within the run object
* the output is being submitted for.
*/
tool_call_id?: string;
}
}
export declare namespace Runs {
export import RequiredActionFunctionToolCall = RunsAPI.RequiredActionFunctionToolCall;
export import Run = RunsAPI.Run;
export import RunStatus = RunsAPI.RunStatus;
export import RunsPage = RunsAPI.RunsPage;
export import RunCreateParams = RunsAPI.RunCreateParams;
export import RunCreateParamsNonStreaming = RunsAPI.RunCreateParamsNonStreaming;
export import RunCreateParamsStreaming = RunsAPI.RunCreateParamsStreaming;
export import RunUpdateParams = RunsAPI.RunUpdateParams;
export import RunListParams = RunsAPI.RunListParams;
export import RunCreateAndPollParams = RunsAPI.RunCreateAndPollParams;
export import RunCreateAndStreamParams = RunsAPI.RunCreateAndStreamParams;
export import RunStreamParams = RunsAPI.RunStreamParams;
export import RunSubmitToolOutputsParams = RunsAPI.RunSubmitToolOutputsParams;
export import RunSubmitToolOutputsParamsNonStreaming = RunsAPI.RunSubmitToolOutputsParamsNonStreaming;
export import RunSubmitToolOutputsParamsStreaming = RunsAPI.RunSubmitToolOutputsParamsStreaming;
export import RunSubmitToolOutputsAndPollParams = RunsAPI.RunSubmitToolOutputsAndPollParams;
export import RunSubmitToolOutputsStreamParams = RunsAPI.RunSubmitToolOutputsStreamParams;
export import Steps = StepsAPI.Steps;
export import CodeInterpreterLogs = StepsAPI.CodeInterpreterLogs;
export import CodeInterpreterOutputImage = StepsAPI.CodeInterpreterOutputImage;
export import CodeInterpreterToolCall = StepsAPI.CodeInterpreterToolCall;
export import CodeInterpreterToolCallDelta = StepsAPI.CodeInterpreterToolCallDelta;
export import FileSearchToolCall = StepsAPI.FileSearchToolCall;
export import FileSearchToolCallDelta = StepsAPI.FileSearchToolCallDelta;
export import FunctionToolCall = StepsAPI.FunctionToolCall;
export import FunctionToolCallDelta = StepsAPI.FunctionToolCallDelta;
export import MessageCreationStepDetails = StepsAPI.MessageCreationStepDetails;
export import RunStep = StepsAPI.RunStep;
export import RunStepDelta = StepsAPI.RunStepDelta;
export import RunStepDeltaEvent = StepsAPI.RunStepDeltaEvent;
export import RunStepDeltaMessageDelta = StepsAPI.RunStepDeltaMessageDelta;
export import ToolCall = StepsAPI.ToolCall;
export import ToolCallDelta = StepsAPI.ToolCallDelta;
export import ToolCallDeltaObject = StepsAPI.ToolCallDeltaObject;
export import ToolCallsStepDetails = StepsAPI.ToolCallsStepDetails;
export import RunStepsPage = StepsAPI.RunStepsPage;
export import StepListParams = StepsAPI.StepListParams;
}
//# sourceMappingURL=runs.d.ts.map