Karini AI Documentation
Go Back to Karini AI
  • Introduction
  • Installation
  • Getting Started
  • Organization
  • User Management
    • User Invitations
    • Role Management
  • Model Hub
    • Embeddings Models
    • Large Language Models (LLMs)
  • Prompt Management
    • Prompt Templates
    • Create Prompt
    • Test Prompt
      • Test & Compare
      • Prompt Observability
      • Prompt Runs
    • Agentic Prompts
      • Create Agent Prompt
      • Test Agent Prompt
    • Prompt Task Types
    • Prompt Versions
  • Datasets
  • Recipes
    • QnA Recipe
      • Data Storage Connectors
      • Connector Credential Setup
      • Vector Stores
      • Create Recipe
      • Run Recipe
      • Test Recipe
      • Evaluate Recipe
      • Export Recipe
      • Recipe Runs
      • Recipe Actions
    • Agent Recipe
      • Agent Recipe Configuration
      • Set up Agentic Recipe
      • Test Agentic Recipe
      • Agentic Evaluation
    • Databricks Recipe
  • Copilots
  • Observability
  • Dashboard Overview
    • Statistical Overview
    • Cost & Usage Summary
      • Spend by LLM Endpoint
      • Spend by Generative AI Application
    • Model Endpoints & Datasets Distribution
    • Dataset Dashboard
    • Copilot Dashboard
    • Model Endpoints Dashboard
  • Catalog Schemas
    • Connectors
    • Catalog Schema Import and Publication Process
  • Prompt Optimization Experiments
    • Set up and execute experiment
    • Optimization Insights
  • Generative AI Workshop
    • Agentic RAG
    • Intelligent Document Processing
    • Generative BI Agentic Assistant
  • Release Notes
Powered by GitBook
On this page
  1. Prompt Management
  2. Agentic Prompts

Test Agent Prompt

PreviousCreate Agent PromptNextPrompt Task Types

Last updated 11 months ago

Agent prompt can be tested using various combinations of LLMs and model parameters, and comparing the responses. Refer to for details about prompt testing.

Following Tracing and Observability features give you good insight into the agent prompt processing when a prompt request is being executed:

Prompt Lens:

Prompt lens let you peek behind the scenes as the agent request is being executed. Here, you can inspect the input we are sending to the language models (LLMs) - including system instructions, context, questions, and response guidelines. For agent prompts, you can also review the following information, as the prompt request is being processed.

  • Agent scratch pad: The scratch pad aids in refining prompts, documenting interactions, or brainstorming ideas based on the outputs received from the selected models.

  • Agent response: The agent response refers to the output or action taken by the selected model in response to a user's prompt or query.

  • Tool response: It gives the insights or summaries related to the tools used, performance metrics, or operational status.

  • Trace: You can see the traces for each operation that is executed during the prompt processing. It includes the following:

    • Input:

    • Output:

  • Attributes: These include various parameters and metrics associated with each request. Some of the attributes include:

    • Input Tokens:

    • Completion tokens:

    • Model parameters such as temperature, max tokens etc.

  • The response displays the following statistics as below:

    • Input Tokens: Total number of input tokens in the LLM request. This includes the prompt instructions, system prompt, context and user query.

    • Output Tokens: Total number of output tokens generated by the LLM in response to the prompt request. This number does not exceed the Max Tokens value setup during the prompt testing.

    • LLM Response Time: The amount of time in milliseconds taken by the LLM to generate complete response for the given prompt request.

    • Time to First Token: The time that it takes for the model to produce the first token of the response after receiving the prompt. TTFT is particularly relevant for applications utilizing streaming, where providing immediate feedback is crucial.

After testing and comparing models, choose the best one. Mark the best model with Select as best answer. The same prompt can be saved with select as best run or Save prompt run option provide.

To see prompt runs in detail refer section.

Test & Compare
Prompt Runs