From 8245b96fec37fbcc42bb53793789f8b5fdb0d74d Mon Sep 17 00:00:00 2001 From: Alex Dixon Date: Fri, 17 Apr 2026 11:44:52 -0700 Subject: [PATCH 1/2] Change 'in-memory' to 'in_memory' in prompt cache enums https://developers.openai.com/api/docs/guides/prompt-caching#configure-per-request --- packages/ai/openai/src/Generated.ts | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/packages/ai/openai/src/Generated.ts b/packages/ai/openai/src/Generated.ts index a24c4f35989..e2fd8a325f3 100644 --- a/packages/ai/openai/src/Generated.ts +++ b/packages/ai/openai/src/Generated.ts @@ -2466,7 +2466,7 @@ export class ChatCompletionFunctions extends S.Class("C /** * The retention policy for the prompt cache. Set to `24h` to enable extended prompt caching, which keeps cached prefixes active for longer, up to a maximum of 24 hours. [Learn more](https://platform.openai.com/docs/guides/prompt-caching#prompt-cache-retention). */ -export class CreateChatCompletionRequestPromptCacheRetentionEnum extends S.Literal("in-memory", "24h") {} +export class CreateChatCompletionRequestPromptCacheRetentionEnum extends S.Literal("in_memory", "24h") {} export class CreateChatCompletionRequest extends S.Class("CreateChatCompletionRequest")({ /** @@ -14969,7 +14969,7 @@ export class CreateResponseTruncationEnum extends S.Literal("auto", "disabled") /** * The retention policy for the prompt cache. Set to `24h` to enable extended prompt caching, which keeps cached prefixes active for longer, up to a maximum of 24 hours. [Learn more](https://platform.openai.com/docs/guides/prompt-caching#prompt-cache-retention). */ -export class CreateResponsePromptCacheRetentionEnum extends S.Literal("in-memory", "24h") {} +export class CreateResponsePromptCacheRetentionEnum extends S.Literal("in_memory", "24h") {} export class CreateResponse extends S.Class("CreateResponse")({ "input": S.optionalWith(InputParam, { nullable: true }), @@ -15181,7 +15181,7 @@ export class ResponseTruncationEnum extends S.Literal("auto", "disabled") {} /** * The retention policy for the prompt cache. Set to `24h` to enable extended prompt caching, which keeps cached prefixes active for longer, up to a maximum of 24 hours. [Learn more](https://platform.openai.com/docs/guides/prompt-caching#prompt-cache-retention). */ -export class ResponsePromptCacheRetentionEnum extends S.Literal("in-memory", "24h") {} +export class ResponsePromptCacheRetentionEnum extends S.Literal("in_memory", "24h") {} export class Response extends S.Class("Response")({ /** From 158e9ee943e324ecb30f541cb2ba12d900fc429c Mon Sep 17 00:00:00 2001 From: Alex Dixon Date: Fri, 17 Apr 2026 12:04:17 -0700 Subject: [PATCH 2/2] Change 'in-memory' to 'in_memory' in enums Updated enum values in prompt cache for consistency. --- .changeset/odd-peaches-chew.md | 5 +++++ 1 file changed, 5 insertions(+) create mode 100644 .changeset/odd-peaches-chew.md diff --git a/.changeset/odd-peaches-chew.md b/.changeset/odd-peaches-chew.md new file mode 100644 index 00000000000..8e37d64be0e --- /dev/null +++ b/.changeset/odd-peaches-chew.md @@ -0,0 +1,5 @@ +--- +"@effect/ai-openai": patch +--- + +Change 'in-memory' to 'in_memory' in prompt cache enums