Table of Contents

Class CreateRunRequest

Namespace
OpenAI.Threads
Assembly
OpenAI-DotNet.dll

Create a run on a thread.

public sealed class CreateRunRequest
Inheritance
CreateRunRequest
Inherited Members

Constructors

CreateRunRequest(string, string, string, string, IEnumerable<Message>, IEnumerable<Tool>, IReadOnlyDictionary<string, string>, double?, double?, ReasoningEffort, int?, int?, TruncationStrategy, string, bool?, JsonSchema, ChatResponseFormat)

Constructor.

public CreateRunRequest(string assistantId, string model = null, string instructions = null, string additionalInstructions = null, IEnumerable<Message> additionalMessages = null, IEnumerable<Tool> tools = null, IReadOnlyDictionary<string, string> metadata = null, double? temperature = null, double? topP = null, ReasoningEffort reasoningEffort = (ReasoningEffort)0, int? maxPromptTokens = null, int? maxCompletionTokens = null, TruncationStrategy truncationStrategy = null, string toolChoice = null, bool? parallelToolCalls = null, JsonSchema jsonSchema = null, ChatResponseFormat responseFormat = ChatResponseFormat.Auto)

Parameters

assistantId string

The ID of the assistant used for execution of this run.

model string

The model that the assistant used for this run.

instructions string

The instructions that the assistant used for this run.

additionalInstructions string

Appends additional instructions at the end of the instructions for the run. This is useful for modifying the behavior on a per-run basis without overriding other instructions.

additionalMessages IEnumerable<Message>

Adds additional messages to the thread before creating the run.

tools IEnumerable<Tool>

The list of tools that the assistant used for this run.

metadata IReadOnlyDictionary<string, string>

Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long.

temperature double?

What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. When null the default temperature (1) will be used.

topP double?

An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We generally recommend altering this or temperature but not both.

reasoningEffort ReasoningEffort

Constrains effort on reasoning for reasoning models. Currently supported values are low, medium, and high. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.

maxPromptTokens int?

The maximum number of prompt tokens that may be used over the course of the run. The run will make a best effort to use only the number of prompt tokens specified, across multiple turns of the run. If the run exceeds the number of prompt tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.

maxCompletionTokens int?

The maximum number of completion tokens that may be used over the course of the run. The run will make a best effort to use only the number of completion tokens specified, across multiple turns of the run. If the run exceeds the number of completion tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.

truncationStrategy TruncationStrategy

Controls for how a thread will be truncated prior to the run. Use this to control the initial context window of the run.

toolChoice string

Controls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.

parallelToolCalls bool?

Whether to enable parallel function calling during tool use.

jsonSchema JsonSchema

The JsonSchema to use for structured JSON outputs.
https://platform.openai.com/docs/guides/structured-outputs
https://json-schema.org/overview/what-is-jsonschema

responseFormat ChatResponseFormat

An object specifying the format that the model must output. Setting to Json or JsonSchema enables JSON mode, which guarantees the message the model generates is valid JSON.
Important: When using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

CreateRunRequest(string, string, string, string, IEnumerable<Message>, IEnumerable<Tool>, IReadOnlyDictionary<string, string>, double?, double?, int?, int?, TruncationStrategy, string, bool?, JsonSchema, ChatResponseFormat)

[Obsolete("use new .ctr")]
public CreateRunRequest(string assistantId, string model, string instructions, string additionalInstructions, IEnumerable<Message> additionalMessages, IEnumerable<Tool> tools, IReadOnlyDictionary<string, string> metadata, double? temperature, double? topP, int? maxPromptTokens = null, int? maxCompletionTokens = null, TruncationStrategy truncationStrategy = null, string toolChoice = null, bool? parallelToolCalls = null, JsonSchema jsonSchema = null, ChatResponseFormat responseFormat = ChatResponseFormat.Auto)

Parameters

assistantId string
model string
instructions string
additionalInstructions string
additionalMessages IEnumerable<Message>
tools IEnumerable<Tool>
metadata IReadOnlyDictionary<string, string>
temperature double?
topP double?
maxPromptTokens int?
maxCompletionTokens int?
truncationStrategy TruncationStrategy
toolChoice string
parallelToolCalls bool?
jsonSchema JsonSchema
responseFormat ChatResponseFormat

Properties

AdditionalInstructions

Appends additional instructions at the end of the instructions for the run. This is useful for modifying the behavior on a per-run basis without overriding other instructions.

[JsonPropertyName("additional_instructions")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingDefault)]
public string AdditionalInstructions { get; }

Property Value

string

AdditionalMessages

Adds additional messages to the thread before creating the run.

[JsonPropertyName("additional_messages")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public IReadOnlyList<Message> AdditionalMessages { get; }

Property Value

IReadOnlyList<Message>

AssistantId

The ID of the assistant used for execution of this run.

[JsonPropertyName("assistant_id")]
public string AssistantId { get; }

Property Value

string

Instructions

The instructions that the assistant used for this run.

[JsonPropertyName("instructions")]
public string Instructions { get; }

Property Value

string

MaxCompletionTokens

The maximum number of completion tokens that may be used over the course of the run. The run will make a best effort to use only the number of completion tokens specified, across multiple turns of the run. If the run exceeds the number of completion tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.

[JsonPropertyName("max_completion_tokens")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public int? MaxCompletionTokens { get; }

Property Value

int?

MaxPromptTokens

The maximum number of prompt tokens that may be used over the course of the run. The run will make a best effort to use only the number of prompt tokens specified, across multiple turns of the run. If the run exceeds the number of prompt tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.

[JsonPropertyName("max_prompt_tokens")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public int? MaxPromptTokens { get; }

Property Value

int?

Metadata

Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long.

[JsonPropertyName("metadata")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public IReadOnlyDictionary<string, string> Metadata { get; }

Property Value

IReadOnlyDictionary<string, string>

Model

The model that the assistant used for this run.

[JsonPropertyName("model")]
public string Model { get; }

Property Value

string

ParallelToolCalls

Whether to enable parallel function calling during tool use.

[JsonPropertyName("parallel_tool_calls")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public bool? ParallelToolCalls { get; }

Property Value

bool?

ReasoningEffort

Constrains effort on reasoning for reasoning models. Currently supported values are low, medium, and high. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.

[JsonInclude]
[JsonPropertyName("reasoning_effort")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingDefault)]
public ReasoningEffort ReasoningEffort { get; }

Property Value

ReasoningEffort

ResponseFormat

[JsonIgnore]
public ChatResponseFormat ResponseFormat { get; }

Property Value

ChatResponseFormat

ResponseFormatObject

An object specifying the format that the model must output. Setting to Json or JsonSchema enables JSON mode, which guarantees the message the model generates is valid JSON.

[JsonPropertyName("response_format")]
[JsonConverter(typeof(ResponseFormatConverter))]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public ResponseFormatObject ResponseFormatObject { get; }

Property Value

ResponseFormatObject

Remarks

Important: When using JSON mode you must still instruct the model to produce JSON yourself via some conversation message, for example via your system message. If you don't do this, the model may generate an unending stream of whitespace until the generation reaches the token limit, which may take a lot of time and give the appearance of a "stuck" request. Also note that the message content may be partial (i.e. cut off) if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.

Stream

If true, returns a stream of events that happen during the Run as server-sent events, terminating when the Run enters a terminal state with a 'data: [DONE]' message.

[JsonPropertyName("stream")]
public bool Stream { get; }

Property Value

bool

Temperature

What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. When null the default temperature (1) will be used.

[JsonPropertyName("temperature")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public double? Temperature { get; }

Property Value

double?

ToolChoice

Controls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.

[JsonPropertyName("tool_choice")]
public dynamic ToolChoice { get; }

Property Value

dynamic

Tools

The list of tools that the assistant used for this run.

[JsonPropertyName("tools")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public IReadOnlyList<Tool> Tools { get; }

Property Value

IReadOnlyList<Tool>

TopP

An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We generally recommend altering this or temperature but not both.

[JsonPropertyName("top_p")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public double? TopP { get; }

Property Value

double?

TruncationStrategy

Controls for how a thread will be truncated prior to the run. Use this to control the initial context window of the run.

[JsonPropertyName("truncation_strategy")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public TruncationStrategy TruncationStrategy { get; }

Property Value

TruncationStrategy