Class CreateRunRequest
Create a run on a thread.
public sealed class CreateRunRequest
- Inheritance
-
CreateRunRequest
- Inherited Members
Constructors
CreateRunRequest(string, string, string, string, IEnumerable<Message>, IEnumerable<Tool>, IReadOnlyDictionary<string, string>, double?, double?, ReasoningEffort, int?, int?, TruncationStrategy, string, bool?, JsonSchema, ChatResponseFormat)
Constructor.
public CreateRunRequest(string assistantId, string model = null, string instructions = null, string additionalInstructions = null, IEnumerable<Message> additionalMessages = null, IEnumerable<Tool> tools = null, IReadOnlyDictionary<string, string> metadata = null, double? temperature = null, double? topP = null, ReasoningEffort reasoningEffort = (ReasoningEffort)0, int? maxPromptTokens = null, int? maxCompletionTokens = null, TruncationStrategy truncationStrategy = null, string toolChoice = null, bool? parallelToolCalls = null, JsonSchema jsonSchema = null, ChatResponseFormat responseFormat = ChatResponseFormat.Auto)
Parameters
assistantId
stringThe ID of the assistant used for execution of this run.
model
stringThe model that the assistant used for this run.
instructions
stringThe instructions that the assistant used for this run.
additionalInstructions
stringAppends additional instructions at the end of the instructions for the run. This is useful for modifying the behavior on a per-run basis without overriding other instructions.
additionalMessages
IEnumerable<Message>Adds additional messages to the thread before creating the run.
tools
IEnumerable<Tool>The list of tools that the assistant used for this run.
metadata
IReadOnlyDictionary<string, string>Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long.
temperature
double?What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. When null the default temperature (1) will be used.
topP
double?An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We generally recommend altering this or temperature but not both.
reasoningEffort
ReasoningEffortConstrains effort on reasoning for reasoning models. Currently supported values are low, medium, and high. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
maxPromptTokens
int?The maximum number of prompt tokens that may be used over the course of the run. The run will make a best effort to use only the number of prompt tokens specified, across multiple turns of the run. If the run exceeds the number of prompt tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.
maxCompletionTokens
int?The maximum number of completion tokens that may be used over the course of the run. The run will make a best effort to use only the number of completion tokens specified, across multiple turns of the run. If the run exceeds the number of completion tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.
truncationStrategy
TruncationStrategyControls for how a thread will be truncated prior to the run. Use this to control the initial context window of the run.
toolChoice
stringControls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.
parallelToolCalls
bool?Whether to enable parallel function calling during tool use.
jsonSchema
JsonSchemaThe JsonSchema to use for structured JSON outputs.
https://platform.openai.com/docs/guides/structured-outputs
https://json-schema.org/overview/what-is-jsonschemaresponseFormat
ChatResponseFormatAn object specifying the format that the model must output. Setting to Json or JsonSchema enables JSON mode, which guarantees the message the model generates is valid JSON.
Important: When using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.
CreateRunRequest(string, string, string, string, IEnumerable<Message>, IEnumerable<Tool>, IReadOnlyDictionary<string, string>, double?, double?, int?, int?, TruncationStrategy, string, bool?, JsonSchema, ChatResponseFormat)
[Obsolete("use new .ctr")]
public CreateRunRequest(string assistantId, string model, string instructions, string additionalInstructions, IEnumerable<Message> additionalMessages, IEnumerable<Tool> tools, IReadOnlyDictionary<string, string> metadata, double? temperature, double? topP, int? maxPromptTokens = null, int? maxCompletionTokens = null, TruncationStrategy truncationStrategy = null, string toolChoice = null, bool? parallelToolCalls = null, JsonSchema jsonSchema = null, ChatResponseFormat responseFormat = ChatResponseFormat.Auto)
Parameters
assistantId
stringmodel
stringinstructions
stringadditionalInstructions
stringadditionalMessages
IEnumerable<Message>tools
IEnumerable<Tool>metadata
IReadOnlyDictionary<string, string>temperature
double?topP
double?maxPromptTokens
int?maxCompletionTokens
int?truncationStrategy
TruncationStrategytoolChoice
stringparallelToolCalls
bool?jsonSchema
JsonSchemaresponseFormat
ChatResponseFormat
Properties
AdditionalInstructions
Appends additional instructions at the end of the instructions for the run. This is useful for modifying the behavior on a per-run basis without overriding other instructions.
[JsonPropertyName("additional_instructions")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingDefault)]
public string AdditionalInstructions { get; }
Property Value
AdditionalMessages
Adds additional messages to the thread before creating the run.
[JsonPropertyName("additional_messages")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public IReadOnlyList<Message> AdditionalMessages { get; }
Property Value
AssistantId
The ID of the assistant used for execution of this run.
[JsonPropertyName("assistant_id")]
public string AssistantId { get; }
Property Value
Instructions
The instructions that the assistant used for this run.
[JsonPropertyName("instructions")]
public string Instructions { get; }
Property Value
MaxCompletionTokens
The maximum number of completion tokens that may be used over the course of the run. The run will make a best effort to use only the number of completion tokens specified, across multiple turns of the run. If the run exceeds the number of completion tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.
[JsonPropertyName("max_completion_tokens")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public int? MaxCompletionTokens { get; }
Property Value
- int?
MaxPromptTokens
The maximum number of prompt tokens that may be used over the course of the run. The run will make a best effort to use only the number of prompt tokens specified, across multiple turns of the run. If the run exceeds the number of prompt tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.
[JsonPropertyName("max_prompt_tokens")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public int? MaxPromptTokens { get; }
Property Value
- int?
Metadata
Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long.
[JsonPropertyName("metadata")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public IReadOnlyDictionary<string, string> Metadata { get; }
Property Value
Model
The model that the assistant used for this run.
[JsonPropertyName("model")]
public string Model { get; }
Property Value
ParallelToolCalls
Whether to enable parallel function calling during tool use.
[JsonPropertyName("parallel_tool_calls")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public bool? ParallelToolCalls { get; }
Property Value
- bool?
ReasoningEffort
Constrains effort on reasoning for reasoning models. Currently supported values are low, medium, and high. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
[JsonInclude]
[JsonPropertyName("reasoning_effort")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingDefault)]
public ReasoningEffort ReasoningEffort { get; }
Property Value
ResponseFormat
[JsonIgnore]
public ChatResponseFormat ResponseFormat { get; }
Property Value
ResponseFormatObject
An object specifying the format that the model must output. Setting to Json or JsonSchema enables JSON mode, which guarantees the message the model generates is valid JSON.
[JsonPropertyName("response_format")]
[JsonConverter(typeof(ResponseFormatConverter))]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public ResponseFormatObject ResponseFormatObject { get; }
Property Value
Remarks
Important: When using JSON mode you must still instruct the model to produce JSON yourself via some conversation message, for example via your system message. If you don't do this, the model may generate an unending stream of whitespace until the generation reaches the token limit, which may take a lot of time and give the appearance of a "stuck" request. Also note that the message content may be partial (i.e. cut off) if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.
Stream
If true, returns a stream of events that happen during the Run as server-sent events, terminating when the Run enters a terminal state with a 'data: [DONE]' message.
[JsonPropertyName("stream")]
public bool Stream { get; }
Property Value
Temperature
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. When null the default temperature (1) will be used.
[JsonPropertyName("temperature")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public double? Temperature { get; }
Property Value
ToolChoice
Controls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.
[JsonPropertyName("tool_choice")]
public dynamic ToolChoice { get; }
Property Value
- dynamic
Tools
The list of tools that the assistant used for this run.
[JsonPropertyName("tools")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public IReadOnlyList<Tool> Tools { get; }
Property Value
TopP
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We generally recommend altering this or temperature but not both.
[JsonPropertyName("top_p")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public double? TopP { get; }
Property Value
TruncationStrategy
Controls for how a thread will be truncated prior to the run. Use this to control the initial context window of the run.
[JsonPropertyName("truncation_strategy")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public TruncationStrategy TruncationStrategy { get; }