config-sample.md +136 −113
1# Sample Configuration1# Sample Configuration
2 2
33Use this example configuration as a starting point. It includes most keys Codex reads from `config.toml`, along with defaults and short notes.Use this example configuration as a starting point. It includes most keys Codex reads from `config.toml`, along with default behaviors, recommended values where helpful, and short notes.
4 4
5For explanations and guidance, see:5For explanations and guidance, see:
6 6
7- [Config basics](https://developers.openai.com/codex/config-basic)7- [Config basics](https://developers.openai.com/codex/config-basic)
8- [Advanced Config](https://developers.openai.com/codex/config-advanced)8- [Advanced Config](https://developers.openai.com/codex/config-advanced)
9- [Config Reference](https://developers.openai.com/codex/config-reference)9- [Config Reference](https://developers.openai.com/codex/config-reference)
1010- [Sandbox and approvals](https://developers.openai.com/codex/security#sandbox-and-approvals)- [Sandbox and approvals](https://developers.openai.com/codex/agent-approvals-security#sandbox-and-approvals)
1111- [Managed configuration](https://developers.openai.com/codex/security#managed-configuration)- [Managed configuration](https://developers.openai.com/codex/enterprise/managed-configuration)
12 12
13Use the snippet below as a reference. Copy only the keys and sections you need into `~/.codex/config.toml` (or into a project-scoped `.codex/config.toml`), then adjust values for your setup.13Use the snippet below as a reference. Copy only the keys and sections you need into `~/.codex/config.toml` (or into a project-scoped `.codex/config.toml`), then adjust values for your setup.
14 14
15```toml15```toml
16# Codex example configuration (config.toml)16# Codex example configuration (config.toml)
17#17#
1818# This file lists all keys Codex reads from config.toml, their default values,# This file lists the main keys Codex reads from config.toml, along with default
1919# and concise explanations. Values here mirror the effective defaults compiled# behaviors, recommended examples, and concise explanations. Adjust as needed.
20# into the CLI. Adjust as needed.
21#20#
22# Notes21# Notes
23# - Root keys must appear before tables in TOML.22# - Root keys must appear before tables in TOML.
28# Core Model Selection27# Core Model Selection
29################################################################################28################################################################################
30 29
3130# Primary model used by Codex. Default: "gpt-5.2-codex" on all platforms.# Primary model used by Codex. Recommended example for most users: "gpt-5.4".
3231model = "gpt-5.2-codex"model = "gpt-5.4"
33 32
3433# Default communication style for supported models. Default: "friendly".# Communication style for supported models. Allowed values: none | friendly | pragmatic
3534# Allowed values: none | friendly | pragmatic# personality = "pragmatic"
36# personality = "friendly"
37 35
38# Optional model override for /review. Default: unset (uses current session model).36# Optional model override for /review. Default: unset (uses current session model).
3937# review_model = "gpt-5.2-codex"# review_model = "gpt-5.4"
40 38
41# Provider id selected from [model_providers]. Default: "openai".39# Provider id selected from [model_providers]. Default: "openai".
42model_provider = "openai"40model_provider = "openai"
44# Default OSS provider for --oss sessions. When unset, Codex prompts. Default: unset.42# Default OSS provider for --oss sessions. When unset, Codex prompts. Default: unset.
45# oss_provider = "ollama"43# oss_provider = "ollama"
46 44
4745# Optional manual model metadata. When unset, Codex auto-detects from model.# Preferred service tier. `fast` is honored only when enabled in [features].
4846# Uncomment to force values.# service_tier = "flex" # fast | flex
47
48# Optional manual model metadata. When unset, Codex uses model or preset defaults.
49# model_context_window = 128000 # tokens; default: auto for model49# model_context_window = 128000 # tokens; default: auto for model
5050# model_auto_compact_token_limit = 0 # tokens; unset uses model defaults# model_auto_compact_token_limit = 64000 # tokens; unset uses model defaults
5151# tool_output_token_limit = 10000 # tokens stored per tool output; default: 10000 for gpt-5.2-codex# tool_output_token_limit = 12000 # tokens stored per tool output
52# model_catalog_json = "/absolute/path/to/models.json" # optional startup-only model catalog override52# model_catalog_json = "/absolute/path/to/models.json" # optional startup-only model catalog override
53# background_terminal_max_timeout = 300000 # ms; max empty write_stdin poll window (default 5m)53# background_terminal_max_timeout = 300000 # ms; max empty write_stdin poll window (default 5m)
54# log_dir = "/absolute/path/to/codex-logs" # directory for Codex logs; default: "$CODEX_HOME/log"54# log_dir = "/absolute/path/to/codex-logs" # directory for Codex logs; default: "$CODEX_HOME/log"
58# Reasoning & Verbosity (Responses API capable models)58# Reasoning & Verbosity (Responses API capable models)
59################################################################################59################################################################################
60 60
6161# Reasoning effort: minimal | low | medium | high | xhigh (default: medium; xhigh on gpt-5.2-codex and gpt-5.2)# Reasoning effort: minimal | low | medium | high | xhigh
6262model_reasoning_effort = "medium"# model_reasoning_effort = "medium"
63
64# Optional override used when Codex runs in plan mode: none | minimal | low | medium | high | xhigh
65# plan_mode_reasoning_effort = "high"
63 66
6467# Reasoning summary: auto | concise | detailed | none (default: auto)# Reasoning summary: auto | concise | detailed | none
65# model_reasoning_summary = "auto"68# model_reasoning_summary = "auto"
66 69
6770# Text verbosity for GPT-5 family (Responses API): low | medium | high (default: medium)# Text verbosity for GPT-5 family (Responses API): low | medium | high
68# model_verbosity = "medium"71# model_verbosity = "medium"
69 72
7073# Force enable or disable reasoning summaries for current model# Force enable or disable reasoning summaries for current model.
71# model_supports_reasoning_summaries = true74# model_supports_reasoning_summaries = true
72 75
73################################################################################76################################################################################
77# Additional user instructions are injected before AGENTS.md. Default: unset.80# Additional user instructions are injected before AGENTS.md. Default: unset.
78# developer_instructions = ""81# developer_instructions = ""
79 82
80# (Ignored) Optional legacy base instructions override (prefer AGENTS.md). Default: unset.
81# instructions = ""
82
83# Inline override for the history compaction prompt. Default: unset.83# Inline override for the history compaction prompt. Default: unset.
84# compact_prompt = ""84# compact_prompt = ""
85 85
86# Override the default commit co-author trailer. Set to "" to disable it.
87# commit_attribution = "Jane Doe <jane@example.com>"
88
86# Override built-in base instructions with a file path. Default: unset.89# Override built-in base instructions with a file path. Default: unset.
87# model_instructions_file = "/absolute/or/relative/path/to/instructions.txt"90# model_instructions_file = "/absolute/or/relative/path/to/instructions.txt"
88 91
89# Migration note: experimental_instructions_file was renamed to model_instructions_file (deprecated).
90
91# Load the compact prompt override from a file. Default: unset.92# Load the compact prompt override from a file. Default: unset.
92# experimental_compact_prompt_file = "/absolute/or/relative/path/to/compact_prompt.txt"93# experimental_compact_prompt_file = "/absolute/or/relative/path/to/compact_prompt.txt"
93 94
94# Legacy name for apply_patch_freeform. Default: false
95include_apply_patch_tool = false
96
97################################################################################95################################################################################
98# Notifications96# Notifications
99################################################################################97################################################################################
100 98
101# External notifier program (argv array). When unset: disabled.99# External notifier program (argv array). When unset: disabled.
102100# Example: notify = ["notify-send", "Codex"]# notify = ["notify-send", "Codex"]
103notify = [ ]
104 101
105################################################################################102################################################################################
106# Approval & Sandbox103# Approval & Sandbox
132# Where to persist CLI login credentials: file (default) | keyring | auto129# Where to persist CLI login credentials: file (default) | keyring | auto
133cli_auth_credentials_store = "file"130cli_auth_credentials_store = "file"
134 131
135132# Base URL for ChatGPT auth flow (not OpenAI API). Default:# Base URL for ChatGPT auth flow (not OpenAI API).
136chatgpt_base_url = "https://chatgpt.com/backend-api/"133chatgpt_base_url = "https://chatgpt.com/backend-api/"
137 134
138# Restrict ChatGPT login to a specific workspace id. Default: unset.135# Restrict ChatGPT login to a specific workspace id. Default: unset.
139136# forced_chatgpt_workspace_id = ""# forced_chatgpt_workspace_id = "00000000-0000-0000-0000-000000000000"
140 137
141# Force login mechanism when Codex would normally auto-select. Default: unset.138# Force login mechanism when Codex would normally auto-select. Default: unset.
142# Allowed values: chatgpt | api139# Allowed values: chatgpt | api
201# If you use --yolo or another full access sandbox setting, web search defaults to live.196# If you use --yolo or another full access sandbox setting, web search defaults to live.
202web_search = "cached"197web_search = "cached"
203 198
204################################################################################
205# Profiles (named presets)
206################################################################################
207
208# Active profile name. When unset, no profile is applied.199# Active profile name. When unset, no profile is applied.
209# profile = "default"200# profile = "default"
210 201
202# Suppress the warning shown when under-development feature flags are enabled.
203# suppress_unstable_features_warning = true
204
211################################################################################205################################################################################
212# Agents (multi-agent roles and limits)206# Agents (multi-agent roles and limits)
213################################################################################207################################################################################
214 208
215209# [agents][agents]
216# Maximum concurrently open agent threads. Default: 6210# Maximum concurrently open agent threads. Default: 6
217# max_threads = 6211# max_threads = 6
218# Maximum nested spawn depth. Root session starts at depth 0. Default: 1212# Maximum nested spawn depth. Root session starts at depth 0. Default: 1
221# job_max_runtime_seconds = 1800215# job_max_runtime_seconds = 1800
222 216
223# [agents.reviewer]217# [agents.reviewer]
224218# description = "Find security, correctness, and test risks in code."# description = "Find correctness, security, and test risks in code."
225# config_file = "./agents/reviewer.toml" # relative to the config.toml that defines it219# config_file = "./agents/reviewer.toml" # relative to the config.toml that defines it
220# nickname_candidates = ["Athena", "Ada"]
226 221
227################################################################################222################################################################################
228# Skills (per-skill overrides)223# Skills (per-skill overrides)
233# path = "/path/to/skill/SKILL.md"228# path = "/path/to/skill/SKILL.md"
234# enabled = false229# enabled = false
235 230
236################################################################################
237# Experimental toggles (legacy; prefer [features])
238################################################################################
239
240experimental_use_unified_exec_tool = false
241
242# Include apply_patch via freeform editing path (affects default tool set). Default: false
243experimental_use_freeform_apply_patch = false
244
245################################################################################231################################################################################
246# Sandbox settings (tables)232# Sandbox settings (tables)
247################################################################################233################################################################################
264[shell_environment_policy]250[shell_environment_policy]
265# inherit: all (default) | core | none251# inherit: all (default) | core | none
266inherit = "all"252inherit = "all"
267253# Skip default excludes for names containing KEY/SECRET/TOKEN (case-insensitive). Default: true# Skip default excludes for names containing KEY/SECRET/TOKEN (case-insensitive). Default: false
268254ignore_default_excludes = trueignore_default_excludes = false
269# Case-insensitive glob patterns to remove (e.g., "AWS_*", "AZURE_*"). Default: []255# Case-insensitive glob patterns to remove (e.g., "AWS_*", "AZURE_*"). Default: []
270exclude = []256exclude = []
271# Explicit key/value overrides (always win). Default: {}257# Explicit key/value overrides (always win). Default: {}
275# Experimental: run via user shell profile. Default: false261# Experimental: run via user shell profile. Default: false
276experimental_use_profile = false262experimental_use_profile = false
277 263
264################################################################################
265# Managed network proxy settings
266################################################################################
267
268[permissions.network]
269# enabled = true
270# proxy_url = "http://127.0.0.1:43128"
271# admin_url = "http://127.0.0.1:43129"
272# enable_socks5 = false
273# socks_url = "http://127.0.0.1:43130"
274# enable_socks5_udp = false
275# allow_upstream_proxy = false
276# dangerously_allow_non_loopback_proxy = false
277# dangerously_allow_non_loopback_admin = false
278# dangerously_allow_all_unix_sockets = false
279# mode = "limited" # limited | full
280# allowed_domains = ["api.openai.com"]
281# denied_domains = ["example.com"]
282# allow_unix_sockets = ["/var/run/docker.sock"]
283# allow_local_binding = false
284
278################################################################################285################################################################################
279# History (table)286# History (table)
280################################################################################287################################################################################
283# save-all (default) | none290# save-all (default) | none
284persistence = "save-all"291persistence = "save-all"
285# Maximum bytes for history file; oldest entries are trimmed when exceeded. Example: 5242880292# Maximum bytes for history file; oldest entries are trimmed when exceeded. Example: 5242880
286293# max_bytes = 0# max_bytes = 5242880
287 294
288################################################################################295################################################################################
289# UI, Notifications, and Misc (tables)296# UI, Notifications, and Misc (tables)
315# You can also add custom .tmTheme files under $CODEX_HOME/themes.322# You can also add custom .tmTheme files under $CODEX_HOME/themes.
316# theme = "catppuccin-mocha"323# theme = "catppuccin-mocha"
317 324
325# Internal tooltip state keyed by model slug. Usually managed by Codex.
326# [tui.model_availability_nux]
327# "gpt-5.4" = 1
328
329# Enable or disable analytics for this machine. When unset, Codex uses its default behavior.
330[analytics]
331enabled = true
332
318# Control whether users can submit feedback from `/feedback`. Default: true333# Control whether users can submit feedback from `/feedback`. Default: true
319[feedback]334[feedback]
320enabled = true335enabled = true
328# "hide_gpt-5.1-codex-max_migration_prompt" = true343# "hide_gpt-5.1-codex-max_migration_prompt" = true
329# model_migrations = { "gpt-4.1" = "gpt-5.1" }344# model_migrations = { "gpt-4.1" = "gpt-5.1" }
330 345
331# Suppress the warning shown when under-development feature flags are enabled.
332# suppress_unstable_features_warning = true
333
334################################################################################346################################################################################
335# Centralized Feature Flags (preferred)347# Centralized Feature Flags (preferred)
336################################################################################348################################################################################
340# shell_tool = true352# shell_tool = true
341# apps = false353# apps = false
342# apps_mcp_gateway = false354# apps_mcp_gateway = false
343# web_search_cached = false
344# web_search_request = false
345# unified_exec = false355# unified_exec = false
346# shell_snapshot = false356# shell_snapshot = false
347# apply_patch_freeform = false
348# multi_agent = false357# multi_agent = false
349# search_tool = false
350# personality = true358# personality = true
351# request_rule = true
352# collaboration_modes = true
353# use_linux_sandbox_bwrap = false359# use_linux_sandbox_bwrap = false
354360# remote_models = false# runtime_metrics = true
355# runtime_metrics = false
356# powershell_utf8 = true361# powershell_utf8 = true
357# child_agents_md = false362# child_agents_md = false
363# sqlite = true
364# fast_mode = true
365# enable_request_compression = true
366# image_generation = false
367# skill_mcp_dependency_install = true
368# skill_env_var_dependency_prompt = false
369# default_mode_request_user_input = false
370# artifact = false
371# prevent_idle_sleep = false
372# responses_websockets = false
373# responses_websockets_v2 = false
374# image_detail_original = false
358 375
359################################################################################376################################################################################
360# Define MCP servers under this table. Leave empty to disable.377# Define MCP servers under this table. Leave empty to disable.
376# tool_timeout_sec = 60.0 # optional; default 60.0 seconds393# tool_timeout_sec = 60.0 # optional; default 60.0 seconds
377# enabled_tools = ["search", "summarize"] # optional allow-list394# enabled_tools = ["search", "summarize"] # optional allow-list
378# disabled_tools = ["slow-tool"] # optional deny-list (applied after allow-list)395# disabled_tools = ["slow-tool"] # optional deny-list (applied after allow-list)
396# scopes = ["read:docs"] # optional OAuth scopes
397# oauth_resource = "https://docs.example.com/" # optional OAuth resource
379 398
380# --- Example: Streamable HTTP transport ---399# --- Example: Streamable HTTP transport ---
381# [mcp_servers.github]400# [mcp_servers.github]
388# startup_timeout_sec = 10.0 # optional407# startup_timeout_sec = 10.0 # optional
389# tool_timeout_sec = 60.0 # optional408# tool_timeout_sec = 60.0 # optional
390# enabled_tools = ["list_issues"] # optional allow-list409# enabled_tools = ["list_issues"] # optional allow-list
410# disabled_tools = ["delete_issue"] # optional deny-list
411# scopes = ["repo"] # optional OAuth scopes
391 412
392################################################################################413################################################################################
393# Model Providers414# Model Providers
394################################################################################415################################################################################
395 416
396# Built-ins include:417# Built-ins include:
397418# - openai (Responses API; requires login or OPENAI_API_KEY via auth flow)# - openai
398419# - oss (Chat Completions API; defaults to http://localhost:11434/v1)# - ollama
420# - lmstudio
399 421
400[model_providers]422[model_providers]
401 423
403# [model_providers.openaidr]425# [model_providers.openaidr]
404# name = "OpenAI Data Residency"426# name = "OpenAI Data Residency"
405# base_url = "https://us.api.openai.com/v1" # example with 'us' domain prefix427# base_url = "https://us.api.openai.com/v1" # example with 'us' domain prefix
406428# wire_api = "responses" # "responses" | "chat" (default varies)# wire_api = "responses" # only supported value
407# # requires_openai_auth = true # built-in OpenAI defaults to true429# # requires_openai_auth = true # built-in OpenAI defaults to true
408# # request_max_retries = 4 # default 4; max 100430# # request_max_retries = 4 # default 4; max 100
409# # stream_max_retries = 5 # default 5; max 100431# # stream_max_retries = 5 # default 5; max 100
410# # stream_idle_timeout_ms = 300000 # default 300_000 (5m)432# # stream_idle_timeout_ms = 300000 # default 300_000 (5m)
433# # supports_websockets = true # optional
411# # experimental_bearer_token = "sk-example" # optional dev-only direct bearer token434# # experimental_bearer_token = "sk-example" # optional dev-only direct bearer token
412# # http_headers = { "X-Example" = "value" }435# # http_headers = { "X-Example" = "value" }
413# # env_http_headers = { "OpenAI-Organization" = "OPENAI_ORGANIZATION", "OpenAI-Project" = "OPENAI_PROJECT" }436# # env_http_headers = { "OpenAI-Organization" = "OPENAI_ORGANIZATION", "OpenAI-Project" = "OPENAI_PROJECT" }
414 437
415438# --- Example: Azure (Chat/Responses depending on endpoint) ---# --- Example: Azure/OpenAI-compatible provider ---
416# [model_providers.azure]439# [model_providers.azure]
417# name = "Azure"440# name = "Azure"
418# base_url = "https://YOUR_PROJECT_NAME.openai.azure.com/openai"441# base_url = "https://YOUR_PROJECT_NAME.openai.azure.com/openai"
419442# wire_api = "responses" # or "chat" per endpoint# wire_api = "responses"
420# query_params = { api-version = "2025-04-01-preview" }443# query_params = { api-version = "2025-04-01-preview" }
421# env_key = "AZURE_OPENAI_API_KEY"444# env_key = "AZURE_OPENAI_API_KEY"
422445# # env_key_instructions = "Set AZURE_OPENAI_API_KEY in your environment"# env_key_instructions = "Set AZURE_OPENAI_API_KEY in your environment"
446# # supports_websockets = false
423 447
424# --- Example: Local OSS (e.g., Ollama-compatible) ---448# --- Example: Local OSS (e.g., Ollama-compatible) ---
425# [model_providers.ollama]449# [model_providers.ollama]
426# name = "Ollama"450# name = "Ollama"
427# base_url = "http://localhost:11434/v1"451# base_url = "http://localhost:11434/v1"
428452# wire_api = "chat"# wire_api = "responses"
429
430################################################################################
431# Profiles (named presets)
432################################################################################
433
434[profiles]
435
436# [profiles.default]
437# model = "gpt-5.2-codex"
438# model_provider = "openai"
439# approval_policy = "on-request"
440# sandbox_mode = "read-only"
441# oss_provider = "ollama"
442# model_reasoning_effort = "medium"
443# model_reasoning_summary = "auto"
444# model_verbosity = "medium"
445# personality = "friendly" # or "pragmatic" or "none"
446# chatgpt_base_url = "https://chatgpt.com/backend-api/"
447# model_catalog_json = "./models.json"
448# experimental_compact_prompt_file = "./compact_prompt.txt"
449# include_apply_patch_tool = false
450# experimental_use_unified_exec_tool = false
451# experimental_use_freeform_apply_patch = false
452# tools.web_search = false # deprecated legacy alias; prefer top-level `web_search`
453# features = { unified_exec = false }
454 453
455################################################################################454################################################################################
456# Apps / Connectors455# Apps / Connectors
474# enabled = false473# enabled = false
475# approval_mode = "approve"474# approval_mode = "approve"
476 475
476################################################################################
477# Profiles (named presets)
478################################################################################
479
480[profiles]
481
482# [profiles.default]
483# model = "gpt-5.4"
484# model_provider = "openai"
485# approval_policy = "on-request"
486# sandbox_mode = "read-only"
487# service_tier = "flex"
488# oss_provider = "ollama"
489# model_reasoning_effort = "medium"
490# plan_mode_reasoning_effort = "high"
491# model_reasoning_summary = "auto"
492# model_verbosity = "medium"
493# personality = "pragmatic" # or "friendly" or "none"
494# chatgpt_base_url = "https://chatgpt.com/backend-api/"
495# model_catalog_json = "./models.json"
496# model_instructions_file = "/absolute/or/relative/path/to/instructions.txt"
497# experimental_compact_prompt_file = "./compact_prompt.txt"
498# tools_view_image = true
499# features = { unified_exec = false }
500
477################################################################################501################################################################################
478# Projects (trust levels)502# Projects (trust levels)
479################################################################################503################################################################################
480 504
481# Mark specific worktrees as trusted or untrusted.
482[projects]505[projects]
506# Mark specific worktrees as trusted or untrusted.
483# [projects."/absolute/path/to/project"]507# [projects."/absolute/path/to/project"]
484# trust_level = "trusted" # or "untrusted"508# trust_level = "trusted" # or "untrusted"
485 509
510################################################################################
511# Tools
512################################################################################
513
514[tools]
515# view_image = true
516
486################################################################################517################################################################################
487# OpenTelemetry (OTEL) - disabled by default518# OpenTelemetry (OTEL) - disabled by default
488################################################################################519################################################################################
496exporter = "none"527exporter = "none"
497# Trace exporter: none (default) | otlp-http | otlp-grpc528# Trace exporter: none (default) | otlp-http | otlp-grpc
498trace_exporter = "none"529trace_exporter = "none"
530# Metrics exporter: none | statsig | otlp-http | otlp-grpc
531metrics_exporter = "statsig"
499 532
500# Example OTLP/HTTP exporter configuration533# Example OTLP/HTTP exporter configuration
501# [otel.exporter."otlp-http"]534# [otel.exporter."otlp-http"]
505# [otel.exporter."otlp-http".headers]538# [otel.exporter."otlp-http".headers]
506# "x-otlp-api-key" = "${OTLP_TOKEN}"539# "x-otlp-api-key" = "${OTLP_TOKEN}"
507 540
508# Example OTLP/gRPC exporter configuration
509# [otel.exporter."otlp-grpc"]
510# endpoint = "https://otel.example.com:4317",
511# headers = { "x-otlp-meta" = "abc123" }
512
513# Example OTLP exporter with mutual TLS
514# [otel.exporter."otlp-http"]
515# endpoint = "https://otel.example.com/v1/logs"
516# protocol = "binary"
517
518# [otel.exporter."otlp-http".headers]
519# "x-otlp-api-key" = "${OTLP_TOKEN}"
520
521# [otel.exporter."otlp-http".tls]541# [otel.exporter."otlp-http".tls]
522# ca-certificate = "certs/otel-ca.pem"542# ca-certificate = "certs/otel-ca.pem"
523# client-certificate = "/etc/codex/certs/client.pem"543# client-certificate = "/etc/codex/certs/client.pem"
524# client-private-key = "/etc/codex/certs/client-key.pem"544# client-private-key = "/etc/codex/certs/client-key.pem"
525```
526 545
527546################################################################################# Example OTLP/gRPC trace exporter configuration
547# [otel.trace_exporter."otlp-grpc"]
548# endpoint = "https://otel.example.com:4317"
549# headers = { "x-otlp-meta" = "abc123" }
528 550
551################################################################################
529# Windows552# Windows
530
531################################################################################553################################################################################
532 554
533[windows]555[windows]
534
535# Native Windows sandbox mode (Windows only): unelevated | elevated556# Native Windows sandbox mode (Windows only): unelevated | elevated
536
537sandbox = "unelevated"557sandbox = "unelevated"
558```