The AI Configuration panel is the central place to manage every aspect of Studio AI: which agents are active, which language models they use, prompt customizations, token usage, MCP server connections, and more. This article covers the panel's layout and each of its tabs.
Click the AI Configuration icon at the far right of the AI Chat panel bar (the second icon from the right). The panel opens alongside the chat, replacing the chat view with the configuration interface.
You can also access it via View > AI Configuration from the menu bar.
The AI Configuration panel is organized into seven tabs along the top:

The Agents tab lists every agent available in Studio AI. For each agent you can:
Changes to agent settings take effect immediately and no restart required.
The Variables tab shows all context variables that are available in the chat via the # prefix. This includes both built-in variables (such as #currentRelativeFilePath) and any custom variables your organization has defined.
For custom variables, you can set or update their values here. Variables defined in this panel are injected into agent prompts automatically when referenced.
The MCP Servers tab lets you manage connections to external tools and data sources through the Model Context Protocol. Studio AI supports both local stdio-based and remote HTTP/SSE MCP servers.
To add a new server, click Add MCP Server and fill in the connection details:
stdio for a local process or HTTP/SSE for a remote endpointOnce connected, the tools provided by the MCP server become available to agents automatically. You can edit or remove any server using the action buttons in the list.
For full details, see MCP Servers.
The Token Usage tab shows a summary table of all LLM usage accumulated in the current Studio session, broken down by model:
Use this table to understand which models are consuming the most tokens, and to spot unexpectedly large requests (for example, attaching very large files as context).
For the full history of individual requests and responses, use the AI Agent History panel — see AI History and Token Usage.
The Prompt Fragments tab lists reusable prompt snippet files (.prompttemplate files) that can be referenced across multiple agent prompt templates. Creating a fragment here makes it available to any agent prompt that imports it.
For details on writing and using prompt fragments, see Customizing Agent Prompts.
The Tools tab shows all the function tools registered with Studio AI — these are the capabilities agents can call during a conversation, such as reading files, running commands, or querying the project structure. You can review which tools are active and, where applicable, enable or disable individual tools.
The Model Aliases tab is where you define the named LLM identifiers that agents use. Instead of hardcoding a specific model name in every agent configuration, agents reference an alias such as default/code or default/universal, and this tab defines what model (or ordered list of models) each alias maps to.
Each alias entry contains:
default/code)Default aliases in Studio AI:
To change which model backs an alias, edit the priority list. This is the recommended way to swap LLM providers across all agents at once without updating each agent individually.
The AI Configuration panel is the central place to manage every aspect of Studio AI: which agents are active, which language models they use, prompt customizations, token usage, MCP server connections, and more. This article covers the panel's layout and each of its tabs.
Click the AI Configuration icon at the far right of the AI Chat panel bar (the second icon from the right). The panel opens alongside the chat, replacing the chat view with the configuration interface.
You can also access it via View > AI Configuration from the menu bar.
The AI Configuration panel is organized into seven tabs along the top:

The Agents tab lists every agent available in Studio AI. For each agent you can:
Changes to agent settings take effect immediately and no restart required.
The Variables tab shows all context variables that are available in the chat via the # prefix. This includes both built-in variables (such as #currentRelativeFilePath) and any custom variables your organization has defined.
For custom variables, you can set or update their values here. Variables defined in this panel are injected into agent prompts automatically when referenced.
The MCP Servers tab lets you manage connections to external tools and data sources through the Model Context Protocol. Studio AI supports both local stdio-based and remote HTTP/SSE MCP servers.
To add a new server, click Add MCP Server and fill in the connection details:
stdio for a local process or HTTP/SSE for a remote endpointOnce connected, the tools provided by the MCP server become available to agents automatically. You can edit or remove any server using the action buttons in the list.
For full details, see MCP Servers.
The Token Usage tab shows a summary table of all LLM usage accumulated in the current Studio session, broken down by model:
Use this table to understand which models are consuming the most tokens, and to spot unexpectedly large requests (for example, attaching very large files as context).
For the full history of individual requests and responses, use the AI Agent History panel — see AI History and Token Usage.
The Prompt Fragments tab lists reusable prompt snippet files (.prompttemplate files) that can be referenced across multiple agent prompt templates. Creating a fragment here makes it available to any agent prompt that imports it.
For details on writing and using prompt fragments, see Customizing Agent Prompts.
The Tools tab shows all the function tools registered with Studio AI — these are the capabilities agents can call during a conversation, such as reading files, running commands, or querying the project structure. You can review which tools are active and, where applicable, enable or disable individual tools.
The Model Aliases tab is where you define the named LLM identifiers that agents use. Instead of hardcoding a specific model name in every agent configuration, agents reference an alias such as default/code or default/universal, and this tab defines what model (or ordered list of models) each alias maps to.
Each alias entry contains:
default/code)Default aliases in Studio AI:
To change which model backs an alias, edit the priority list. This is the recommended way to swap LLM providers across all agents at once without updating each agent individually.