Class CreateThreadAndRunRequest
public sealed class CreateThreadAndRunRequest
- Inheritance
-
CreateThreadAndRunRequest
- Inherited Members
Constructors
CreateThreadAndRunRequest(string, string, string, IReadOnlyList<Tool>, ToolResources, IReadOnlyDictionary<string, string>, double?, double?, ReasoningEffort, int?, int?, TruncationStrategy, string, bool?, JsonSchema, ChatResponseFormat, CreateThreadRequest)
Constructor.
public CreateThreadAndRunRequest(string assistantId, string model = null, string instructions = null, IReadOnlyList<Tool> tools = null, ToolResources toolResources = null, IReadOnlyDictionary<string, string> metadata = null, double? temperature = null, double? topP = null, ReasoningEffort reasoningEffort = (ReasoningEffort)0, int? maxPromptTokens = null, int? maxCompletionTokens = null, TruncationStrategy truncationStrategy = null, string toolChoice = null, bool? parallelToolCalls = null, JsonSchema jsonSchema = null, ChatResponseFormat responseFormat = ChatResponseFormat.Auto, CreateThreadRequest createThreadRequest = null)
Parameters
assistantId
stringThe ID of the assistant to use to execute this run.
model
stringThe ID of the Model to be used to execute this run. If a value is provided here, it will override the model associated with the assistant. If not, the model associated with the assistant will be used.
instructions
stringOverride the default system message of the assistant. This is useful for modifying the behavior on a per-run basis.
tools
IReadOnlyList<Tool>Override the tools the assistant can use for this run. This is useful for modifying the behavior on a per-run basis.
toolResources
ToolResourcesA set of resources that are used by the assistant's tools. The resources are specific to the type of tool. For example, the 'code_interpreter' tool requires a list of file IDs, while the 'file_search' tool requires a list of vector store IDs.
metadata
IReadOnlyDictionary<string, string>Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long.
temperature
double?What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. When null the default temperature (1) will be used.
topP
double?An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We generally recommend altering this or temperature but not both.
reasoningEffort
ReasoningEffortConstrains effort on reasoning for reasoning models. Currently supported values are low, medium, and high. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
maxPromptTokens
int?The maximum number of prompt tokens that may be used over the course of the run. The run will make a best effort to use only the number of prompt tokens specified, across multiple turns of the run. If the run exceeds the number of prompt tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.
maxCompletionTokens
int?The maximum number of completion tokens that may be used over the course of the run. The run will make a best effort to use only the number of completion tokens specified, across multiple turns of the run. If the run exceeds the number of completion tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.
truncationStrategy
TruncationStrategyControls for how a thread will be truncated prior to the run. Use this to control the initial context window of the run.
toolChoice
stringControls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.
parallelToolCalls
bool?Whether to enable parallel function calling during tool use.
jsonSchema
JsonSchemaThe JsonSchema to use for structured JSON outputs.
https://platform.openai.com/docs/guides/structured-outputs
https://json-schema.org/overview/what-is-jsonschemaresponseFormat
ChatResponseFormatAn object specifying the format that the model must output. Setting to Json or JsonSchema enables JSON mode, which guarantees the message the model generates is valid JSON.
Important: When using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.createThreadRequest
CreateThreadRequestOptional, CreateThreadRequest.
CreateThreadAndRunRequest(string, string, string, IReadOnlyList<Tool>, ToolResources, IReadOnlyDictionary<string, string>, double?, double?, int?, int?, TruncationStrategy, string, bool?, JsonSchema, ChatResponseFormat, CreateThreadRequest)
[Obsolete("use new .ctr")]
public CreateThreadAndRunRequest(string assistantId, string model, string instructions, IReadOnlyList<Tool> tools, ToolResources toolResources, IReadOnlyDictionary<string, string> metadata, double? temperature, double? topP, int? maxPromptTokens, int? maxCompletionTokens, TruncationStrategy truncationStrategy, string toolChoice, bool? parallelToolCalls, JsonSchema jsonSchema, ChatResponseFormat responseFormat = ChatResponseFormat.Auto, CreateThreadRequest createThreadRequest = null)
Parameters
assistantId
stringmodel
stringinstructions
stringtools
IReadOnlyList<Tool>toolResources
ToolResourcesmetadata
IReadOnlyDictionary<string, string>temperature
double?topP
double?maxPromptTokens
int?maxCompletionTokens
int?truncationStrategy
TruncationStrategytoolChoice
stringparallelToolCalls
bool?jsonSchema
JsonSchemaresponseFormat
ChatResponseFormatcreateThreadRequest
CreateThreadRequest
Properties
AssistantId
The ID of the assistant to use to execute this run.
[JsonPropertyName("assistant_id")]
public string AssistantId { get; }
Property Value
Instructions
Override the default system message of the assistant. This is useful for modifying the behavior on a per-run basis.
[JsonPropertyName("instructions")]
public string Instructions { get; }
Property Value
MaxCompletionTokens
The maximum number of completion tokens that may be used over the course of the run. The run will make a best effort to use only the number of completion tokens specified, across multiple turns of the run. If the run exceeds the number of completion tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.
[JsonPropertyName("max_completion_tokens")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public int? MaxCompletionTokens { get; }
Property Value
- int?
MaxPromptTokens
The maximum number of prompt tokens that may be used over the course of the run. The run will make a best effort to use only the number of prompt tokens specified, across multiple turns of the run. If the run exceeds the number of prompt tokens specified, the run will end with status 'incomplete'. See 'incomplete_details' for more info.
[JsonPropertyName("max_prompt_tokens")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public int? MaxPromptTokens { get; }
Property Value
- int?
Metadata
Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format. Keys can be a maximum of 64 characters long and values can be a maximum of 512 characters long.
[JsonPropertyName("metadata")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public IReadOnlyDictionary<string, string> Metadata { get; }
Property Value
Model
The ID of the Model to be used to execute this run. If a value is provided here, it will override the model associated with the assistant. If not, the model associated with the assistant will be used.
[JsonPropertyName("model")]
public string Model { get; }
Property Value
ParallelToolCalls
Whether to enable parallel function calling during tool use.
[JsonPropertyName("parallel_tool_calls")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public bool? ParallelToolCalls { get; }
Property Value
- bool?
ReasoningEffort
Constrains effort on reasoning for reasoning models. Currently supported values are low, medium, and high. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
[JsonInclude]
[JsonPropertyName("reasoning_effort")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingDefault)]
public ReasoningEffort ReasoningEffort { get; }
Property Value
ResponseFormat
[JsonIgnore]
public ChatResponseFormat ResponseFormat { get; }
Property Value
ResponseFormatObject
An object specifying the format that the model must output. Setting to Json or JsonSchema enables JSON mode, which guarantees the message the model generates is valid JSON.
[JsonPropertyName("response_format")]
[JsonConverter(typeof(ResponseFormatConverter))]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public ResponseFormatObject ResponseFormatObject { get; }
Property Value
Remarks
Important: When using JSON mode you must still instruct the model to produce JSON yourself via some conversation message, for example via your system message. If you don't do this, the model may generate an unending stream of whitespace until the generation reaches the token limit, which may take a lot of time and give the appearance of a "stuck" request. Also note that the message content may be partial (i.e. cut off) if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.
Stream
If true, returns a stream of events that happen during the Run as server-sent events, terminating when the Run enters a terminal state with a 'data: [DONE]' message.
[JsonPropertyName("stream")]
public bool Stream { get; }
Property Value
Temperature
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. When null the default temperature (1) will be used.
[JsonPropertyName("temperature")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public double? Temperature { get; }
Property Value
ThreadRequest
The optional CreateThreadRequest options to use.
[JsonPropertyName("thread")]
public CreateThreadRequest ThreadRequest { get; }
Property Value
ToolChoice
Controls which (if any) tool is called by the model. none means the model will not call any tools and instead generates a message. auto is the default value and means the model can pick between generating a message or calling one or more tools. required means the model must call one or more tools before responding to the user. Specifying a particular tool like {"type": "file_search"} or {"type": "function", "function": {"name": "my_function"}} forces the model to call that tool.
[JsonPropertyName("tool_choice")]
public dynamic ToolChoice { get; }
Property Value
- dynamic
ToolResources
A set of resources that are used by the assistant's tools. The resources are specific to the type of tool. For example, the 'code_interpreter' tool requires a list of file IDs, while the 'file_search' tool requires a list of vector store IDs.
[JsonPropertyName("tool_resources")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public ToolResources ToolResources { get; }
Property Value
Tools
The list of tools that the assistant used for this run.
[JsonPropertyName("tools")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public IReadOnlyList<Tool> Tools { get; }
Property Value
TopP
An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered. We generally recommend altering this or temperature but not both.
[JsonPropertyName("top_p")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public double? TopP { get; }
Property Value
TruncationStrategy
Controls for how a thread will be truncated prior to the run. Use this to control the initial context window of the run.
[JsonPropertyName("truncation_strategy")]
[JsonIgnore(Condition = JsonIgnoreCondition.WhenWritingNull)]
public TruncationStrategy TruncationStrategy { get; }