Skip to main content

Models

Introduction

DIAL allows you to access and use AI models from all major LLM providers, open-source community, alternative vendors, and fine-tuned micro models, as well as self-hosted or models listed on Hugging Face, NVIDIA NIM or DeepSeek.

Refer to Supported Models to learn more about the supported AI models.

DIAL provides a single Unified API, based on OpenAI API, for accessing all AI models, embedding models and applications. The key design principle is to create a unification layer that allows all models and applications to be interchangeable, delivering a cohesive conversational experience and future-proof development of custom GenAI applications.

DIAL uses AI model adapters, that unify the APIs of respective AI models to align with the Unified Protocol of DIAL Core. If a model is compatible with DIAL Unified Protocol, it can be used directly without an adapter.

You can add AI models to DIAL via a direct configuration of DIAL Core or using DIAL Admin Panel. This tutorial describes how to add and manage AI models using DIAL Admin Panel.

Main Screen

On this page, you can find all AI models deployments available on your DIAL instance. Here you can view, filter, and add new model definitions.

When a model deployment is successfully added, it can be accessed across the DIAL ecosystem including, DIAL Chat, DIAL Marketplace and via API for authorized user roles.

You can use DIAL Core API to access available model deployments and to send chat completion requests.

Models grid

TIP: Click Columns to open the columns selector to define what columns to display and in which order.

The grid with models displays the main properties of models which include:

FieldDescription
Display NameModel name (e.g. "GPT-4 Turbo"). Display name is shown in all DIAL client UI components for a quick model identification.
VersionVersion of a specific model deployment (e.g. 0613, v1). Used to distinguish between "latest", "beta" or date-stamped builds.
DescriptionDescription of the model's purpose including any relevant details. The description is displayed in DIAL Chat UI and Marketplace.
IDUnique key under the models section of DIAL Core's config. Must match the upstream service's model or deployment name (e.g. gpt-4-0613).
Source TypeSource type of a model:
Adapter: Model is based on a model adapter.
External Endpoint: Model is deployed outside DIAL infrastructure and exposes an endpoint DIAL Core uses for communication.
Model Container: Model is based on a deployed model container.
SourceSource identifier. Adapter ID, Model serving container ID or Endpoint of the model, based on your Source Type selection.
AuthorInformation about the user who deployed the model.
TypeDefines Chat (conversational completions) and Embedding models (vector generation). DIAL Core uses this to choose the correct API endpoint and a payload schema.
Override NameContext-specific optional display label that overrides the Display Name in UI components. Use it to give a model different aliases in different workflows without redefining the model.
TopicsTags or categories you can assign for discovery, filtering, or grouping in large deployments (e.g. "finance", "support", "image-capable"). Helps end users and admins find the right model by the use case. Topics are also used to filter models in DIAL Marketplace.
Attachment typesTypes of attachments this model can accept according to MIME types.
Max attachment numberMaximum number of attachments allowed in a single request. Blank for an unlimited number.
Tokenizer modelIdentifies the specific model with a tokenization algorithm identical to the referenced model's. This is typically the name of the earliest released model in a series of models sharing an identical tokenization algorithm. This parameter is essential for DIAL clients that reimplement tokenization algorithms on their side, instead of utilizing the tokenize endpoint provided by the model.
Interaction limitThe interaction limit parameter in models refers to the maximum number of tokens that can be transmitted in a completion request and response combined. This parameter ensures that the model does not exceed a specified token limit during interactions.
Prompt priceCost per unit (according to Cost unit, typically "token" or "request") applied to the input portion of each call. Used by the Dashboard and Usage Logs to estimate spending in real time.
Completion priceCost per unit is charged for the output portion of each call. Combined with the prompt price, it determines your per-model cost calculations.

Create Model

Follow these steps to add new AI model to your DIAL instance:

  1. Click + Create to invoke a Create Model modal.

  2. Define parameters for a new model:

    FieldRequiredDescription
    IDYesUnique identifier used by the model adapter to invoke the model's backend.
    Display NameYesModel name shown across the UI (e.g. "GPT-4 Turbo").
    Display versionNoVersion is an optional tag to track releases when you register multiple variants of the same model. (e.g. 2024-07-18, v1)
    DescriptionNoFree-text note about the model's purpose or distinguishing traits.
    Source typeYesAdapter: Select the corresponding AI model adapter from the list of available adapters. In this case DIAL Core will use the adapter's endpoint URL to communicate with the model.
    Model Container: Select one of the available model containers deployed in DIAL. In this case DIAL Core will use the container URL to communicate with the model.
    External Endpoint: For externally-hosted models, provide the chat completion endpoint URL DIAL Core will use to directly (not using model adapters) communicate with the model. In this case, the model API must be compatible with DIAL Core API.
  3. Click Create to close the dialog and open the configuration screen. When done with model configuration, click Save. It may take some time for the changes to take effect after saving. Once added, the model appears in the Models listing and become available to use across the DIAL ecosystem.

Configuration

You can access the model configuration screen by clicking any model in the models grid and also when adding a new model. In this section, you can view and configure all settings for the selected language model deployment.

  • Properties: Main definitions and runtime settings.
  • Features: Optional capabilities and custom endpoints.
  • Roles: User groups that can invoke this model and their rate limits.
  • Interceptors: Custom logic to modify requests or responses.
  • Audit: Provides aggregated audit metrics and detailed logs of individual configuration changes.

Properties

In the Properties tab, you can view and edit main definitions and runtime settings for model deployment.

  • Basic Information: Identification and basic information about the model and its author.
  • Source Type: Parameters associated with the way the model was created: Adapter, Model Container, or External Endpoint.
  • Personalization: Parameters to customize model's appearance on UI.
  • Attachments: Define attachment types and maximum number of attachments the model can accept.
  • Default Parameters: Set default values for model parameters used in chat/completions requests.
  • Upstream Configuration: Define upstream endpoints, authentication keys, weights, and extra data.
  • Advanced Options: Tokenizer model, forward auth token, interaction limits, retry attempts.
  • Cost Configuration: Set cost unit, prompt price, and completion price for real-time billing.
Basic Information
FieldRequiredEditableDescription
ID-NoUnique key DIAL Core uses in the models section. Must match the upstream's deployment or model name (e.g. gpt-4o, gpt-4-turbo). Non-editable after the model created.
Updated Time-NoDate and time when the model's configuration was last updated.
Creation Time-NoDate and time when the model's configuration was created.
Sync with core-NoIndicates the state of the entity's configuration synchronization between Admin and DIAL Core.
Synchronization occurs automatically every 2 mins (configurable via CONFIG_AUTO_RELOAD_SCHEDULE_DELAY_MILLISECONDS).
Important: Sync state is not available for sensitive information (API keys/tokens/auth settings).
Synced:
Entity's states are identical in Admin and in Core for valid entities or entity is missing in Core for invalid entities.
In progress...:
If Synced conditions are not met and changes were applied within last 2 mins (this period is configurable via CONFIG_EXPORT_SYNC_DURATION_THRESHOLD_MS).
Out of sync:
If Synced conditions are not met and changes were applied more than 2 mins ago (this period is configurable via CONFIG_EXPORT_SYNC_DURATION_THRESHOLD_MS).
Unavailable:
Displayed when it is not possible to determine the entity's state in Core. This occurs if:
- The config was not received from Core for any reason.
- The configuration of entities in Core is not entirely compatible with the one in the Admin service.
Display NameYesYesModel name shown in tables and dropdowns in DIAL clients (e.g. "GPT-4o"). Helps users identify and select models on UI.
Display versionNoYesVersion tag for tracking releases (e.g. 0613, v1). Useful for A/B testing or canary rollouts.
DescriptionNoYesText describing the model's features, details, or other relevant information.
MaintainerNoYesName of the responsible person or team overseeing the model's configuration.
Source typeYesYesSource type indicates the way a model was created: Adapter, Model Container, External Endpoint.

Source Type

The Source Type defines how the model is integrated into DIAL. Depending on the selected Source Type, different parameters need to be specified.

Adapter

Unless AI model API is compatible with the DIAL Unified Protocol, you need adapters to be able to use AI models in DIAL. Model adapters unify the APIs of respective AI models to align with the Unified Protocol of DIAL Core. If the Source Type of your model is Adapter, DIAL Core will use adapters endpoint to communicate with the model.

DIAL includes adapters for Azure OpenAI models, GCP Vertex AI models, and AWS Bedrock models. You can also create custom adapters for other AI models with DIAL SDK.

Adapter can be added in Builders/Adapters.

The following properties need to be specified if selected Source Type is Adapter:

FieldRequiredDescription
AdapterYesModel adapter that will be used to handle requests to this model deployment (e.g. OpenAI, DIAL). Adapter defines how to authenticate, format payloads, and parse responses.
TypeYesSelect Chat or Embedding API.
Chat: Conversational chat completions.
Embedding: Vector generation (semantic search, clustering).
EndpointYesURL that DIAL Core will invoke for this model. The base URL is determined by the selected adapter, while the path can be partially customized.

Model Container

AI models can be deployed in DIAL using images. You can deploy images and use them to create containers in the Deployments section.

If the Source Type of your model is Model Container, DIAL Core will use the container URL to communicate with the model.

The following properties need to be specified if selected Source Type is Model Container:

FieldRequiredDescription
TypeYesSelect Chat or Embedding API.
Chat: Conversational chat completions.
Embedding: Vector generation (semantic search, clustering).
ContainerYesSelect one of the running Model Containers.
EndpointYesURL that DIAL Core will invoke for this model. The base URL is determined by the selected container, while the path can be partially customized.

External Endpoint

If your AI model is deployed elsewhere and is compatible with DIAL Core API, you can add it using its chat completion endpoint for a direct communication between the DIAL Core and the AI model.

The following properties need to be specified if selected Source Type is External Endpoint:

FieldRequiredDescription
TypeYesSelect Chat or Embedding API.
Chat: Conversational chat completions.
Embedding: Vector generation (semantic search, clustering).
EndpointYesURL that DIAL Core will invoke for this model.

Personalization

These parameters help customize how the model is presented in the DIAL UI.

FieldRequiredDescription
Override nameNoCustom display name for specific contexts.
IconNoLogo to visually distinguish models in the UI.
TopicsNoTag that associates a model with one or more topics or categories (e.g. "finance", "support").

Attachments
FieldRequiredDescription
AttachmentsNoAttachment types (images, files) this model can accept.
None: Attachments are not allowed.
All: Unrestricted types. Optionally specify max number of attachments.
Custom: Specific MIME types.
Attachments max numberNoMaximum number of input attachments, When unset (default), attachments are not allowed.

Default Parameters

Default parameters can be use to pass additional parameters with the chat completion request.

Upstream Configuration
FieldDescription
Upstream EndpointsOne or more backend URLs to send requests to. Enables round-robin load balancing or fallback among multiple hosts. Refer to Load Balancer to learn more.
You can use upstream endpoint to provide an alternative URL. For example a model Docker container URL if the model is deployed as a container in DIAL. Refer to Model Servings to learn more about this use case.
KeysAPI key, token, or credential passed to the upstream. Stored securely and masked—click the eye icon to reveal.
WeightNumeric weight for this endpoint in a multi-upstream scenario. Higher = more traffic share.
TierSpecifies an endpoint group. In a regular scenario, all requests are routed to endpoints with the lowest tier, but in case of an outage or hitting the limits, the next one in the line helps to handle the load.
Extra DataFree-form JSON or string metadata passed to the model adapter with each request.

Advanced Options
FieldDescription
Tokenizer ModelIdentifies the specific model whose tokenization algorithm exactly matches that of the referenced model. This is typically the name of the earliest released model in a series of models sharing an identical tokenization algorithm. This parameter is essential for DIAL clients that reimplement tokenization algorithms on their side, instead of utilizing the tokenize endpoint provided by the model.
Forward auth tokenSelect a downstream auth token to forward from the user’s session (for downstream multi-tenant).
Interaction limitThis parameter ensures that the model does not exceed a specified token limit during interactions.
Available values:
None: DIAL does not apply any additional interaction limits beyond limits that your model enforces natively. Ideal for early prototyping or when you trust the LLM’s built-in safeguards.
Total number of tokens: Enforces a single, cumulative cap on the sum of all prompt + completion tokens across the entire chat.
Separately Prompts & Completions: Two independent limits: one on the sum of all input (prompt) tokens and another on the sum of all output (completion) tokens over the course of a conversation.
Max retry attemptsThe number of times DIAL Core will retry a connection in case of upstream errors (e.g. on timeouts or 5xx responses).
Hashing OrderSpecifies the order in which parts of a chat completion request—like the tools used and the chat messages — are combined to create a unique fingerprint (hash) for that request. By default, it checks the tools first and then the messages. This fingerprint helps the system recognize when different requests share the same beginning, so it can reuse previous work and speed things up through caching, which is especially important because the system can only cache information for requests that start the same way. Refer to DIAL Core documentation for details.

Cost Configuration

Enables real-time cost estimation and quota enforcement. Powers the telemetry dashboard with per-model spending metrics.

FieldRequiredDescription
Cost unitYesBase unit for billing.
Available values:
None: Disables all cost tracking for this model.
Tokens: Every token sent or received by the model is counted towards your cost metrics.
Char without whitespace: Tells DIAL to count only non-whitespace characters (letters, numbers, punctuation) in each request as the billing unit.
Prompt priceYesCost per unit for prompt tokens.
Completion priceYesCost per unit for completion tokens (chat responses).

Features

In the Features tab, you can enable, disable, or override optional capabilities for a specific model. You can use model's features to tailor DIAL Core’s Unified Protocol behavior—turning features on when your model supports them, or off when it doesn’t.

Model features apply per AI model, controlling what the model endpoint itself supports (e.g. whether GPT-4 can accept system prompts or function calls).

TIP: Enable only the features you need. Extra toggles can cause errors if upstream doesn’t support them. After setting a custom endpoint, test it via a simple API call to confirm accessibility and authentication.

Custom Endpoints

Some models adapters expose specialized HTTP endpoints for tokenization, rate estimation, prompt truncation, or live configuration. You can override the default Unified Protocol calls by specifying them in this section.

FieldDescription
Rate endpointURL to invoke the model’s cost‐estimation or billing API.
Call an endpoint that returns token counts & credit usage.
Override if your adapter supports a dedicated "rate" path.
Tokenize endpointURL to invoke a standalone tokenization service.
Use when you need precise token counts before truncation or batching.
Models without built-in tokenization require this.
Truncate prompt endpointURL to invoke a prompt‐truncation API.
Ensures prompts are safely cut to max context length.
Useful when working with very long user inputs.
Configuration endpointURL to fetch JSON Schema describing settings of the DIAL model. DIAL Core exposes this endpoint to DIAL clients as GET v1/deployments/<deployment name>/configuration. DIAL client must provide a JSON value corresponding to the configuration JSON Schema in a chat completion request in the custom_fields.configuration field.
Feature Flags (Toggles)

Model toggles ensure you don’t accidentally send unsupported parameters to a given model.

Each toggle corresponds to a capability in the Unified Protocol. Enable them only if your model and adapter fully support that feature.

ToggleDescription
System promptAllows injecting a system‐level message (the "agent’s instructions") at the start of every chat. Disable for models that ignore or block system prompts.
ToolsEnables the tools (a.k.a. functions) feature for safe external API calls. Enable if you plan to use DIAL toolsets or function calling.
Temperature supportedEnables the temperature parameter to control randomness in model deployment's output.
SeedEnables the seed parameter for deterministic output. If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism is not guaranteed, and you should refer to the system_fingerprint response parameter to monitor changes in the backend.
URL AttachmentsAllows passing URLs as attachments (images, docs) to the model. Can be required for image-based or file-referencing prompts.
Folder AttachmentsEnables attaching folders (batching multiple files).
Assistant attachment in requestIndicates whether the model deployment supports attachments chat completion request messages form the role=assistant. When enabled, DIAL Chat must preserve attachments in messages from assistants, instead of removing them. The feature is especially useful for models that can generate attachments as well as take attachments as an input. A typical example of such a model is an image-editing model.
Accessible by request keyIndicates whether the deployment is accessible using a per-request API key.
Content partsIndicates whether the model deployment supports requests with content parts or not. Content parts in a chat completion request allow sending to AI model a message consisting of different types of content, such as text, images, or other media, within a single request which enables more rich and flexible interaction with a model.
CacheWhether the deployment supports LLM caching.
Auto cachingIndicates whether the deployment supports automatic caching, where it's possible.
Parallel tool callsIndicates whether the deployment supports parallel_tool_calls parameter in a chat completion request which enables parallel function calling when using a tool.
Support comment in rate responseIndicates whether the model supports the field comment in rate response payload.

Roles

You can create and manage roles in the Access Management section or via a direct configuration of DIAL Core.

In the Roles tab, you can define user groups that are authorized to use a specific model deployment and enforce role-based rate and usage limits. This is essential for multi-tenant governance, quota enforcement, and cost control across teams or customers, preventing runaway costs by enforcing a hard ceiling.

Important: if roles are not specified for a specific model deployment, it is available to all users.

  • Refer to Access & Cost Control to learn more about access control in DIAL.
  • Refer to Roles to lean more about roles in DIAL.
  • Refer to tutorials to learn how to use roles to define access and cost control for JWT and API keys

Roles grid
ColumnDescription
IDUnique role's identifier.
Display NameUnique role's name displayed on UI.
DescriptionDescription of the role's purpose (e.g., "DIAL Prompt Engineering Team").
Tokens per minutePer Minute tokens limit for a specific role. Blank = no limits. Inherits the default value. Can be overridden.
Tokens per dayDaily tokens limit for a specific role. Blank = no limits. Inherits the default value. Can be overridden.
Tokens per weekWeekly tokens limit for a specific role. Blank = no limits. Inherits the default value. Can be overridden.
Tokens per monthMonthly tokens limit for a specific role. Blank = no limits. Inherits the default value. Can be overridden.
ActionsAdditional role-specific actions.
When Make available to specific roles toggle is off - opens the Roles section in a new tab.
When Make available to specific roles toggle is on, you can open the Roles section in a new tab, set Set unlimited, Remove the role from the list or Reset to default limits.

Set Rate Limits

The grid on the Roles screen lists the roles that can access a specific model. Here, you can also set individual limits for selected roles. For example, you can give "Admin" role unlimited monthly tokens but throttle "Developer" to 100,000 tokens/day or allow the "External Partner" role a small trial quota (e.g., 10,000 tokens/month) before upgrade.

To set or change rate limits for a role
  1. Click in the desired cell (e.g., Tokens per day for the "ADMIN").
  2. Enter a numeric limit or leave blank to enable an unlimited access. Click Reset to default limits to restore default settings for all roles.
  3. Click Save to apply changes.

Default Limits

Default limits are set for all roles in the Roles grid by default; however you can override them as needed.

FieldDescription
Default tokens per minuteThe maximum tokens any user can consume per minute unless a specific limit is in place.
Default tokens per dayThe maximum tokens any user can consume per day unless a specific limit is in place.
Default tokens per weekThe maximum tokens any user can consume per week unless a specific limit is in place.
Default tokens per monthThe maximum tokens any user may consume per month unless a specific limit is in place.

Role-Specific Access

Use Make available to specific roles toggle to define access to the model:

  • Off: Model is callable by any authenticated user. All existing user roles are in the grid.
  • On: Model is restricted - only the roles you explicitly add to the grid can invoke it.

Add

You can add a role only if Make available to specific roles toggle is On.

  1. Click + Add (top-right of the Roles Grid).
  2. Select one or more roles in the modal. The list or roles is defined in the Access Management section.
  3. Confirm to add role(s) to the table.

Remove

You can remove a role only if Make available to specific roles toggle is On.

  1. Click the actions menu in the role's line.
  2. Choose Remove in the menu.

Interceptors

DIAL uses Interceptors to add custom logic to in/out requests for models and apps, enabling PII obfuscation, guardrails, safety checks, and beyond.

Refer to Interceptors to learn more.

In the Interceptors tab, you can view configured global interceptors and define local interceptors that will be triggered for the selected model. You can define Interceptors in the Entities → Interceptors section to add them to the processing pipeline of DIAL Core.

Interceptors grid
ColumnDescription
IDUnique interceptor identifier.
OrderExecution sequence. Interceptors run in ascending order (1 → 2 → 3...). A request will flow through each interceptor's in this order.Response interceptors are invoked in the reversed order.
Display NameAlias of the interceptor, matching the Name field in its definition.
DescriptionFree-text summary from the interceptor's definition, explaining its purpose.
ActionsAdditional actions.
Open interceptor in a new tab.
Remove the selected interceptor from the model's configuration.

Add

Follow these steps to attach one or more interceptors to the model's configuration:

  1. Click + Add (in the upper-right of the interceptors grid).
  2. In the Add Interceptors modal, choose one or more from the grid of defined interceptors.
  3. Apply to append them to the bottom of the list (are added in the same order as selected in the modal).

TIP: If you need a new interceptor, first create it under Entities → Interceptors and then revisit this tab to attach it to the model's configuration.

Reorder

  1. Drag & Drop the handle (⋮⋮⋮⋮) to reassign the order in which interceptors are triggered.
  2. Release to reposition; order renumbers automatically.
  3. Save to lock-in the new execution sequence.

Remove

  1. Click the actions menu in the interceptor's row.
  2. Choose Remove to detach it from this model.
  3. Save to lock-in the interceptors list.

Audit

In the Audit tab, you can monitor key metrics, activities and traces related to the selected language model.

Dashboard

TIP: You can monitor the entire system's metrics in Telemetry.

In the Dashboard tab, you can monitor real-time and historical metrics for the model. You can use it to monitor usage patterns, enforce SLAs, optimize costs, and troubleshoot anomalies.

ControlDescription
Time PeriodSelect the date range for all charts and tables (e.g. last 15 min, 2 days, 7 days, 30 days).
+ Add filterFilter with options to drill into a specific project.
Auto refreshSet the dashboard to poll for new data (e.g. every 1 min) or turn off auto-refresh.
System Usage Chart

A time-series line chart of requests throughput over time. You can use it to monitor traffic peaks and valleys, correlate spikes with deployments or feature roll outs.

Key Metrics

Four high-level metrics are displayed alongside the chart. All calculated for the selected time period.

You can use them to:

  • Chargeback to internal teams or external customers by "Money".
  • Track adoption via "Unique Users".
  • Monitor burst traffic with "Request Count".
  • Watch token consumption to anticipate quota exhaustion.
MetricDescription
Unique UsersCount of distinct user IDs or API keys that have called this model.
Request CountTotal number of chat or embedding calls routed to this model.
Total TokensSum of prompt + completion tokens consumed by this model.
MoneyEstimated spending on this model.
Projects Consumption Table

This table shows the KPIs breakdown by Project. You can use it to compare consumption across multiple projects.

ColumnDescription
ProjectThe project utilizing this model.
Request CountNumber of calls directed to the model.
Prompt tokensTotal tokens submitted in the prompt portion of requests.
Completion tokensTotal tokens returned by the model as responses.
MoneyEstimated costs.

Traces

TIP: You can monitor the entire system's traces in Usage Log.

In this tab, you can see individual traces, each representing a single end-to-end interaction of a DIAL entity with the selected model.

ColumnDescription
Completion TimeTimestamp when the trace finished processing (end-to-end interaction).
Trace IDUnique identifier of the trace (one end-to-end interaction).
TopicAuto-generated subject/title summarizing the trace.
ReactionsIndication of user reactions presence (like/dislike) for the trace.
Cached prompt tokensNumber of prompt tokens served from cache (prompt-caching).
Prompt tokensNumber of tokens in the prompt sent to the model for this trace.
Completion tokensNumber of tokens generated by the model as output for this trace.
Deployment priceCost attributed to the selected deployment for this trace.
Total priceTotal cost of the trace.
Number of request messagesNumber of discrete request messages that were included in the trace.
Deployment IDIdentifier of the DIAL deployment used to serve this trace.
Parent Deployment IDIdentifier of the parent deployment (e.g., application that was using the underlying model).
ModelIdentifier of the underlying model used to carry out the trace.
ProjectProject to which this trace associated in DIAL.
UpstreamThe upstream endpoint (e.g., completions endpoint of the model).
Execution pathExecution path of the trace.
UserIdentifier of the end user who initiated the trace.
User titleName of the user (if available).
LanguageLanguage detected in the trace (e.g., en).
DurationTotal end-to-end duration of the trace from first request to completion.
Response IDIdentifier of the response object returned by the model for this trace.
Conversation IDIdentifier of the conversation/session this trace belongs to.
Code span IDIdentifier of a specific code execution span associated with the trace (if any).
Code span parent IDIdentifier of the parent span for a code execution span (if any).

Conversations

TIP: You can monitor all usage sessions in Usage Log.

In Conversations, you can see individual traces grouped into end‑to‑end conversation sessions.

ColumnDescription
Last activityTimestamp of the most recent trace within the conversation.
Conversation IDUnique identifier of the user session that groups related traces.
TopicAuto-generated subject summarizing the conversation.
Cached prompt tokensCount of prompt tokens served from cache across the conversation.
Prompt tokensTotal number of request/prompt tokens sent to the model across all traces in the conversation.
Completion tokensTotal number of tokens generated by the model across all traces in the conversation.
Total priceAggregated cost for the conversation.
Number of request messagesTotal number of discrete request messages included in the conversation.
Deployment IDIdentifier of the deployment associated with the conversation.
ProjectProject to which the conversation associated in DIAL.
UserIdentifier of the end user who initiated the conversation.
User titleName of the user (if available).
LanguageDetected language for the conversation (e.g., en).

Activities

The Activities section provides detailed visibility into all changes made to the selected model. This section mimics the functionality available in the global Audit → Activities menu, but is scoped specifically to the selected model.

85.png

Activities List Table
FieldDescription
Activity typeAction performed on the model (e.g., Create, Update, Delete).
TimeTimestamp indicating when the activity occurred.
InitiatedEmail address of the user who performed the activity.
Activity IDUnique identifier for the logged activity, used for tracking and auditing.
ActionsAvailable actions:
- View details: Click to open a new screen with activity details. Refer to Activity Details to learn more.
- Resource rollback: click to restore a previous version. Refer to Resource Rollback for details.
Activity Details

The Activity Details view provides a detailed snapshot of a specific change made to a model.

86.png

To open Activity Details, click on the three-dot menu (⋮) at the end of a row in the Activities grid and select “View Details”.

Element/SectionDescription
Activity typeType of the change performed (e.g., Update, Create, Delete).
TimeTimestamp of the change.
InitiatedIdentifier of the user who made the change.
Activity IDUnique identifier for the specific activity tracking.
ComparisonDropdown to switch between showing all parameter or changed only.
ViewDropdown to switch for selection between Before/After and Before/Current state.
Parameters DiffSide-by-side comparison of model fields values before and after the change. Color-coding is used to indicate the operation type (Update, Create, Delete).
Resource Rollback

Use Resource Rollback to restore the previous version of the selected activity. A rollback leads to generation of a new entry on the audit activity screen.

JSON Editor

Advanced users with technical expertise can work with the model properties in a JSON editor view mode. It is useful for advanced scenarios of bulk updates, copy/paste between environments, or tweaking settings not exposed on UI.

TIP: You can switch between UI and JSON only if there are no unsaved changes.

In JSON editor, you can use the view dropdown to select between Admin format and Core format. Note, that these formatting options are for your convenience only and do not render properties as they are defined in DIAL Core. After making changes, the Sync with core indicator on the main configuration screen will inform you about the synchronization state with DIAL Core.

Working with JSON Editor
  1. Navigate to Entities → Models, then select the model you want to edit.
  2. Click the JSON Editor toggle (top-right). The UI reveals the raw JSON.
  3. Chose between the Admin and Core format to see and work with properties in the necessary format. Note: Core format view mode does not render the actual configuration stored in DIAL Core but the configuration in Admin service displayed in the DIAL Core format.
  4. Make changes and click Save to apply them.
  5. After making changes, the Sync with core indicator on the main configuration screen will inform you about the synchronization state with DIAL Core.

Delete

Use the Delete button in the Configuration screen toolbar to permanently remove the selected language model.