AI Configuration: Managing Agents and Settings

The AI Configuration panel is the central place to manage every aspect of Studio AI: which agents are active, which language models they use, prompt customizations, token usage, MCP server connections, and more. This article covers the panel's layout and each of its tabs.

Opening AI Configuration

Click the AI Configuration icon at the far right of the AI Chat panel bar (the second icon from the right). The panel opens alongside the chat, replacing the chat view with the configuration interface.

You can also access it via View > AI Configuration from the menu bar.

Panel Overview

The AI Configuration panel is organized into seven tabs along the top:

Tab Purpose
Agents Enable or disable agents, assign language models, configure notifications
Variables View and set global context variables available to all agents
MCP Servers Add, edit, and remove Model Context Protocol server connections
Token Usage Monitor cumulative token consumption by model
Prompt Fragments Manage reusable prompt snippet files
Tools View and manage the tools (functions) available to agents
Model Aliases Define named LLM aliases and priority lists used across agents

The AI Configuration panel showing seven tabs across the top (Agents, Variables, MCP Servers, Token Usage, Prompt Fragments, Tools, Model Aliases) with the Agents tab active and a list of agents with enable toggles and model selectors.

Agents Tab

The Agents tab lists every agent available in Studio AI. For each agent you can:

  • Enable or disable the agent using the toggle. A disabled agent does not appear in the chat and cannot be addressed.
  • Set the language model: choose which model (or model alias) the agent uses by default. If your organization has configured multiple LLMs, you can assign different models to different agents. For example, a faster model for @Universal and a more capable model for @Coder.
  • Configure completion notifications: toggle whether the agent sends a notification when it completes a task. This is particularly useful for long-running Agent Mode sessions.

Changes to agent settings take effect immediately and no restart required.

Variables Tab

The Variables tab shows all context variables that are available in the chat via the # prefix. This includes both built-in variables (such as #currentRelativeFilePath) and any custom variables your organization has defined.

For custom variables, you can set or update their values here. Variables defined in this panel are injected into agent prompts automatically when referenced.

MCP Servers Tab

The MCP Servers tab lets you manage connections to external tools and data sources through the Model Context Protocol. Studio AI supports both local stdio-based and remote HTTP/SSE MCP servers.

To add a new server, click Add MCP Server and fill in the connection details:

  • Name - a display name for the server
  • Type - stdio for a local process or HTTP/SSE for a remote endpoint
  • Command / URL - the command to run (stdio) or the endpoint URL (HTTP/SSE)
  • Arguments and environment variables - any additional parameters the server requires

Once connected, the tools provided by the MCP server become available to agents automatically. You can edit or remove any server using the action buttons in the list.

For full details, see MCP Servers.

Token Usage Tab

The Token Usage tab shows a summary table of all LLM usage accumulated in the current Studio session, broken down by model:

Column Description
Model The LLM that handled the requests
Input Tokens Total tokens sent to the model (prompts, context, files)
Output Tokens Total tokens generated by the model (responses)
Total Tokens Combined input + output
Last Used Timestamp of the most recent request to this model

Use this table to understand which models are consuming the most tokens, and to spot unexpectedly large requests (for example, attaching very large files as context).

For the full history of individual requests and responses, use the AI Agent History panel — see AI History and Token Usage.

Prompt Fragments Tab

The Prompt Fragments tab lists reusable prompt snippet files (.prompttemplate files) that can be referenced across multiple agent prompt templates. Creating a fragment here makes it available to any agent prompt that imports it.

For details on writing and using prompt fragments, see Customizing Agent Prompts.

Tools Tab

The Tools tab shows all the function tools registered with Studio AI — these are the capabilities agents can call during a conversation, such as reading files, running commands, or querying the project structure. You can review which tools are active and, where applicable, enable or disable individual tools.

Model Aliases Tab

The Model Aliases tab is where you define the named LLM identifiers that agents use. Instead of hardcoding a specific model name in every agent configuration, agents reference an alias such as default/code or default/universal, and this tab defines what model (or ordered list of models) each alias maps to.

Each alias entry contains:

  • Alias name - the identifier referenced in agent configurations (e.g., default/code)
  • Priority list - an ordered list of LLM provider/model pairs. Studio AI uses the first available model in the list; if it is unavailable, it falls through to the next

Default aliases in Studio AI:

Alias Typical use
default/code @Coder, @ClaudeCode, and other code-focused agents
default/universal @Universal and general question-answering agents
default/fabric K2View-specific agents (e.g., @LU, @Interfaces, @Broadway-Edit)
default/summarize Session naming, summary, and background agents
default/code-completion Inline code completion

To change which model backs an alias, edit the priority list. This is the recommended way to swap LLM providers across all agents at once without updating each agent individually.

AI Configuration: Managing Agents and Settings

The AI Configuration panel is the central place to manage every aspect of Studio AI: which agents are active, which language models they use, prompt customizations, token usage, MCP server connections, and more. This article covers the panel's layout and each of its tabs.

Opening AI Configuration

Click the AI Configuration icon at the far right of the AI Chat panel bar (the second icon from the right). The panel opens alongside the chat, replacing the chat view with the configuration interface.

You can also access it via View > AI Configuration from the menu bar.

Panel Overview

The AI Configuration panel is organized into seven tabs along the top:

Tab Purpose
Agents Enable or disable agents, assign language models, configure notifications
Variables View and set global context variables available to all agents
MCP Servers Add, edit, and remove Model Context Protocol server connections
Token Usage Monitor cumulative token consumption by model
Prompt Fragments Manage reusable prompt snippet files
Tools View and manage the tools (functions) available to agents
Model Aliases Define named LLM aliases and priority lists used across agents

The AI Configuration panel showing seven tabs across the top (Agents, Variables, MCP Servers, Token Usage, Prompt Fragments, Tools, Model Aliases) with the Agents tab active and a list of agents with enable toggles and model selectors.

Agents Tab

The Agents tab lists every agent available in Studio AI. For each agent you can:

  • Enable or disable the agent using the toggle. A disabled agent does not appear in the chat and cannot be addressed.
  • Set the language model: choose which model (or model alias) the agent uses by default. If your organization has configured multiple LLMs, you can assign different models to different agents. For example, a faster model for @Universal and a more capable model for @Coder.
  • Configure completion notifications: toggle whether the agent sends a notification when it completes a task. This is particularly useful for long-running Agent Mode sessions.

Changes to agent settings take effect immediately and no restart required.

Variables Tab

The Variables tab shows all context variables that are available in the chat via the # prefix. This includes both built-in variables (such as #currentRelativeFilePath) and any custom variables your organization has defined.

For custom variables, you can set or update their values here. Variables defined in this panel are injected into agent prompts automatically when referenced.

MCP Servers Tab

The MCP Servers tab lets you manage connections to external tools and data sources through the Model Context Protocol. Studio AI supports both local stdio-based and remote HTTP/SSE MCP servers.

To add a new server, click Add MCP Server and fill in the connection details:

  • Name - a display name for the server
  • Type - stdio for a local process or HTTP/SSE for a remote endpoint
  • Command / URL - the command to run (stdio) or the endpoint URL (HTTP/SSE)
  • Arguments and environment variables - any additional parameters the server requires

Once connected, the tools provided by the MCP server become available to agents automatically. You can edit or remove any server using the action buttons in the list.

For full details, see MCP Servers.

Token Usage Tab

The Token Usage tab shows a summary table of all LLM usage accumulated in the current Studio session, broken down by model:

Column Description
Model The LLM that handled the requests
Input Tokens Total tokens sent to the model (prompts, context, files)
Output Tokens Total tokens generated by the model (responses)
Total Tokens Combined input + output
Last Used Timestamp of the most recent request to this model

Use this table to understand which models are consuming the most tokens, and to spot unexpectedly large requests (for example, attaching very large files as context).

For the full history of individual requests and responses, use the AI Agent History panel — see AI History and Token Usage.

Prompt Fragments Tab

The Prompt Fragments tab lists reusable prompt snippet files (.prompttemplate files) that can be referenced across multiple agent prompt templates. Creating a fragment here makes it available to any agent prompt that imports it.

For details on writing and using prompt fragments, see Customizing Agent Prompts.

Tools Tab

The Tools tab shows all the function tools registered with Studio AI — these are the capabilities agents can call during a conversation, such as reading files, running commands, or querying the project structure. You can review which tools are active and, where applicable, enable or disable individual tools.

Model Aliases Tab

The Model Aliases tab is where you define the named LLM identifiers that agents use. Instead of hardcoding a specific model name in every agent configuration, agents reference an alias such as default/code or default/universal, and this tab defines what model (or ordered list of models) each alias maps to.

Each alias entry contains:

  • Alias name - the identifier referenced in agent configurations (e.g., default/code)
  • Priority list - an ordered list of LLM provider/model pairs. Studio AI uses the first available model in the list; if it is unavailable, it falls through to the next

Default aliases in Studio AI:

Alias Typical use
default/code @Coder, @ClaudeCode, and other code-focused agents
default/universal @Universal and general question-answering agents
default/fabric K2View-specific agents (e.g., @LU, @Interfaces, @Broadway-Edit)
default/summarize Session naming, summary, and background agents
default/code-completion Inline code completion

To change which model backs an alias, edit the priority list. This is the recommended way to swap LLM providers across all agents at once without updating each agent individually.