Karini AI Documentation
Go Back to Karini AI
  • Introduction
  • Installation
  • Getting Started
  • Organization
  • User Management
    • User Invitations
    • Role Management
  • Model Hub
    • Embeddings Models
    • Large Language Models (LLMs)
  • Prompt Management
    • Prompt Templates
    • Create Prompt
    • Test Prompt
      • Test & Compare
      • Prompt Observability
      • Prompt Runs
    • Agentic Prompts
      • Create Agent Prompt
      • Test Agent Prompt
    • Prompt Task Types
    • Prompt Versions
  • Datasets
  • Recipes
    • QnA Recipe
      • Data Storage Connectors
      • Connector Credential Setup
      • Vector Stores
      • Create Recipe
      • Run Recipe
      • Test Recipe
      • Evaluate Recipe
      • Export Recipe
      • Recipe Runs
      • Recipe Actions
    • Agent Recipe
      • Agent Recipe Configuration
      • Set up Agentic Recipe
      • Test Agentic Recipe
      • Agentic Evaluation
    • Databricks Recipe
  • Copilots
  • Observability
  • Dashboard Overview
    • Statistical Overview
    • Cost & Usage Summary
      • Spend by LLM Endpoint
      • Spend by Generative AI Application
    • Model Endpoints & Datasets Distribution
    • Dataset Dashboard
    • Copilot Dashboard
    • Model Endpoints Dashboard
  • Catalog Schemas
    • Connectors
    • Catalog Schema Import and Publication Process
  • Prompt Optimization Experiments
    • Set up and execute experiment
    • Optimization Insights
  • Generative AI Workshop
    • Agentic RAG
    • Intelligent Document Processing
    • Generative BI Agentic Assistant
  • Release Notes
Powered by GitBook
On this page
  • Answer
  • Prompt Lens
  • Trace
  • Statistics
  1. Recipes
  2. QnA Recipe

Test Recipe

PreviousRun RecipeNextEvaluate Recipe

Last updated 8 months ago

A recipe can be tested by bringing and configuring the element into the recipe canvas (see ).

Click the Test button to open a chat window, allowing interaction through queries. Submit your question review the response generated by the recipe. The response includes the following:

Answer

You can see the real-time response from your recipe RAG pipeline that includes the answer to the question, prompt lens icon, a trace icon, and statistics. If the model selected in the prompt for the recipe supports streaming, you will see a streaming response.

Prompt Lens

Prompt lens lets you peek behind the scenes as the request is being executed. Here, you can inspect the input sent to the language models (LLMs) - including system instructions, context, questions, and prompt. This empowers you to analyze the quality of your retrieved context from the vector store and make necessary adjustments to the strategy if needed.

Trace

Click on the trace icon to view detailed step-by-step information about the prompt request processing and response generation. Trace has two sections as Prompt and Attributes.

  1. Prompt: Shows the traces of each operation executed during the processing . It includes the following:

    • Input

    • Output

  2. Attributes: Shows various parameters and metrics associated with each request. Some of the attributes include:

    • Input Tokens

    • Completion tokens

    • Model parameters such as temperature, max tokens etc.

Statistics

Following statistics are displayed when the response is generated after a test.

  • Search Embeddings: Time taken in milliseconds to retrieve the similar embeddings based on the user query.

  • Question Embeddings Creation: The time taken in milliseconds to generate embeddings for a given question.

  • LLM Response Time: The amount of time in milliseconds taken by the LLM to generate complete response for the given prompt request.

  • LLM Request Timestamp: Represents the specific time a request was made to the LLM.

  • Time To First Token: The time time taken in milliseconds by the LLM to produce the first token of the response after receiving the prompt. TTFT is particularly relevant for applications utilizing streaming, where providing immediate feedback is crucial.

  • Input Tokens: Total number of input tokens in the LLM request. This includes the prompt instructions, system prompt, context and user query.

  • Embeddings Input Tokens: Number of tokens, converted into vectors by an embedding model.

Output Tokens: Total number of output tokens generated by the LLM in response to the prompt request. This number does not exceed the value setup during the prompt testing.

Input Unsafety Score: It measures the score for the given input. A higher score indicates a greater level of unsafety.

Input Toxicity Score: This score represents the likelihood that the input text could be perceived as or harmful.

Doc Summarization: The amount of time, in milliseconds, taken to summarize the retrieved embedding chunks. This number is reported when you select Summarize chunks option in configuration.

Summarization Prompt Tokens: Number of tokens in the prompt or input provided for a summarization task. This number is reported when you select Summarize chunks option in configuration.

Summarization Response Tokens: Number of tokens in the generated response of a summarization task. This number is reported when you select Summarize chunks option in configuration.

Create Recipe
unsafety
toxic
Context Generation
Context Generation
Context Generation
Output
context generation
Max Tokens