Connectors
The Karini AI platform supports the creation of Catalog schemas, which provide a structured representation of data within a catalog. A Catalog Schema defines the format, and relationships of datasets stored in a database or data management system, ensuring that data is well-structured, accessible, and optimized for querying.
By enabling schema creation with various data connectors, Karini AI allows users to integrate structured data sources, ensuring efficient data retrieval, seamless AI-driven workflows, and scalable data operations for business intelligence and automation.
To create a Catalog schema in the Karini AI platform, navigate to the Catalog Schema section and click Add new to begin the process.
First, enter a user-friendly name for the catalog schema, ensuring it is descriptive and easily identifiable.
Next, choose the appropriate connector type from the available options and provide the required authentication credentials based on the selected connector.
Providing connectors to the leading databases and data warehouses:
Glue: AWS Glue is a fully managed extract, transform, and load (ETL) service that enables seamless data integration and cataloging. It automates schema discovery, data preparation, and metadata management, making it ideal for building scalable data lakes and pipelines. For more details refer to AWS Glue Documentation.
AWS credentials
AWS IAM Role ARN (Required) : Enter the IAM role ARN that grants access to AWS Glue and related services.
AWS External ID (Required) : A unique identifier used for secure cross-account authentication.
AWS Default Region (Required) : Select the AWS region where the Glue catalog and data are hosted (e.g., us-east-1).
Enter S3 Path (Required) : Provide the Amazon S3 path where the dataset associated with the Glue catalog is stored.
MySQL
MySQL credentials
Host (Required) : Enter the hostname or IP address of the MySQL server.
User Name (Required) : Provide the MySQL database username for authentication.
User Password (Required) : Enter the password associated with the MySQL user account.
PostgreSQL
Postgres Credentials
Host (Required) : Enter the hostname or IP address of the PostgreSQL database server.
User Name (Required) : Provide the PostgreSQL database username for authentication.
User Password (Required) : Enter the password associated with the PostgreSQL user account.
Database (Optional) – Specify the database name within the PostgreSQL instance (if applicable).
Port (Optional) – Enter the port number for PostgreSQL connection (default is 5432).
Snowflake
Snowflake credentials
Account (Required) : Provide the Snowflake account identifier used to connect to the database.
User Name (Required) : Enter the Snowflake username for authentication.
User Password (Required) : Provide the password associated with the Snowflake user account.
Redshift
Redshift credentials
Cluster ID (Required) : Provide the unique identifier of the Amazon Redshift cluster.
Database (Required) : Enter the name of the database within the Redshift cluster.
User Name (Required) : Enter the username for authentication to connect to the Redshift database.
User Password (Required) : Provide the password associated with the Redshift user account.
Host (Required) : Provide the Redshift cluster endpoint (hostname or IP).
Port (Required) : Enter the port number on which the Redshift database listens.
MS SQL
MS SQL credentials
Host (Required) : Provide the hostname or IP address of the MS SQL server.
User Name (Required) : Enter the username for authentication to connect to the MS SQL database.
User Password (Required) : Provide the password associated with the MS SQL user account.
Oracle
Oracle credentials
Host (Required) : Provide the hostname or IP address of the Oracle database server.
User Name (Required) : Enter the username for authentication to connect to the Oracle database.
User Password (Required) : Provide the password associated with the Oracle user account.
Service Name (Required) : Specify the Oracle service name used for the database connection.
Port (Required) : Enter the port number on which the Oracle database listens.
Databricks Unity Catalog
Databricks credentials
Databricks Host URL (Required) : Enter the Databricks workspace host URL (e.g., https://<databricks-instance>.cloud.databricks.com).
Databricks API Token (Required) : Provide the Databricks API authentication token for secure access.
Databricks SQL Warehouse HTTP Path (Required) : Enter the HTTP path to the Databricks SQL warehouse for query execution.
Teradata
Teradata credentials
Host (Required) : Provide the hostname or IP address of the Teradata database server.
User Name (Required) : Enter the username for authentication to connect to the Teradata database.
User Password (Required) : Provide the password associated with the Teradata user account.
Database (Optional) : Specify the name of the database within the Teradata system.
After entering the required credentials, proceed to save the catalog schema. Once saved, its status will be displayed as Pending Import.
Last updated