You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Document in the README that both stateful and stateless multi-turn modes work
(Optional, future) If Copilot upstream's state retention turns out to differ from OpenAI's (e.g., shorter TTL, different store semantics), surface that in docs
Acceptance criteria
Two-turn test: turn 1 returns id: \"resp_X\". Turn 2 sends only the new user message + previous_response_id: \"resp_X\". Reply correctly references context from turn 1.
Part of #1. Depends on #3.
Goal
OpenAI's Responses API supports two ways to carry context across turns:
encrypted_content) intoinput(covered by Model-to-endpoint routing (chat vs responses) #5)previous_response_id: \"resp_…\"and only sends the new user turn; the server reconstitutes the prior contextWe should support both. (1) is the path Codex CLI prefers for ZDR; (2) is what some lighter Responses clients use.
Current state
previous_response_idis unmentioned anywhere in the repo (Grep0 hits, see #1 audit).Tasks
previous_response_id?: stringto the request type (Add upstream Responses API service client #3)id: string(the response id) to the response type if not already there from Add upstream Responses API service client #3storesemantics), surface that in docsAcceptance criteria
id: \"resp_X\". Turn 2 sends only the new user message +previous_response_id: \"resp_X\". Reply correctly references context from turn 1.File pointers