Skip to content
Draft
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Remove max_tokens implementation detail from doc comments
Co-authored-by: Copilot <223556219+Copilot@users.noreply.github.com>
  • Loading branch information
MackinnonBuck and Copilot committed Mar 31, 2026
commit 47d07813dbeaaa11764c392751568ecdc5922e95
3 changes: 1 addition & 2 deletions dotnet/src/Types.cs
Original file line number Diff line number Diff line change
Expand Up @@ -1118,8 +1118,7 @@ public class ProviderConfig

/// <summary>
/// Maximum number of tokens the model can generate in a single response.
/// Sent as <c>max_tokens</c> per LLM API request. When hit, the model stops
/// generating and returns a truncated response.
/// When hit, the model stops generating and returns a truncated response.
/// </summary>
[JsonPropertyName("maxOutputTokens")]
public int? MaxOutputTokens { get; set; }
Comment thread
MackinnonBuck marked this conversation as resolved.
Expand Down
3 changes: 1 addition & 2 deletions go/types.go
Original file line number Diff line number Diff line change
Expand Up @@ -602,8 +602,7 @@ type ProviderConfig struct {
// Azure contains Azure-specific options
Azure *AzureProviderOptions `json:"azure,omitempty"`
// MaxOutputTokens is the maximum number of tokens the model can generate in a single response.
// Sent as max_tokens per LLM API request. When hit, the model stops generating and returns
// a truncated response.
// When hit, the model stops generating and returns a truncated response.
MaxOutputTokens int `json:"maxOutputTokens,omitempty"`
// MaxPromptTokens is the maximum number of tokens allowed in the prompt for a single LLM API
// request. Used by the runtime to trigger conversation compaction before sending a request
Expand Down
1 change: 0 additions & 1 deletion nodejs/src/types.ts
Original file line number Diff line number Diff line change
Expand Up @@ -1263,7 +1263,6 @@ export interface ProviderConfig {

/**
* Maximum number of tokens the model can generate in a single response.
* Sent as {@link https://platform.openai.com/docs/api-reference/chat/create#chat-create-max_tokens max_tokens} per LLM API request.
* When hit, the model stops generating and returns a truncated response.
*/
maxOutputTokens?: number;
Expand Down
3 changes: 1 addition & 2 deletions python/copilot/session.py
Original file line number Diff line number Diff line change
Expand Up @@ -508,8 +508,7 @@ class ProviderConfig(TypedDict, total=False):
bearer_token: str
azure: AzureProviderOptions # Azure-specific options
# Maximum number of tokens the model can generate in a single response.
# Sent as max_tokens per LLM API request. When hit, the model stops
# generating and returns a truncated response.
# When hit, the model stops generating and returns a truncated response.
max_output_tokens: int
# Maximum number of tokens allowed in the prompt for a single LLM API request.
# Used by the runtime to trigger conversation compaction before sending a
Expand Down
Loading