Generating SQL for Snowflake using Ollama, Qdrant
This notebook runs through the process of using the vanna
Python package to generate SQL using AI (RAG + LLMs) including connecting to a database and training. If you're not ready to train on your own database, you can still try it using a sample SQLite database.
Run Using Colab Open in GitHub
Which LLM do you want to use?
-
OpenAI via Vanna.AI (Recommended)Use Vanna.AI for free to generate your queries
-
OpenAIUse OpenAI with your own API key
-
Azure OpenAIIf you have OpenAI models deployed on Azure
-
AnthropicUse Anthropics Claude with your Anthropic API Key
-
[Selected] OllamaUse Ollama locally for free. Requires additional setup.
-
Google GeminiUse Google Gemini with your Gemini or Vertex API Key
-
Mistral via Mistral APIIf you have a Mistral API key
-
Other LLMIf you have a different LLM model
Where do you want to store the 'training' data?
-
Vanna Hosted Vector DB (Recommended)Use Vanna.AIs hosted vector database (pgvector) for free. This is usable across machines with no additional setup.
-
ChromaDBUse ChromaDBs open-source vector database for free locally. No additional setup is necessary -- all database files will be created and stored locally.
-
[Selected] QdrantUse Qdrants open-source vector database
-
MarqoUse Marqo locally for free. Requires additional setup. Or use their hosted option.
-
Other VectorDBUse any other vector database. Requires additional setup.
Setup
%pip install 'vanna[qdrant,ollama,snowflake]'
from vanna.ollama import Ollama
from vanna.qdrant import Qdrant_VectorStore
from qdrant_client import QdrantClient
class MyVanna(Qdrant_VectorStore, Ollama):
def __init__(self, config=None):
Qdrant_VectorStore.__init__(self, config=config)
Ollama.__init__(self, config=config)
vn = MyVanna(config={'client': 'QdrantClient(...)', 'model': 'mistral'})
Which database do you want to query?
-
Postgres
-
Microsoft SQL Server
-
MySQL
-
DuckDB
-
[Selected] Snowflake
-
BigQuery
-
SQLite
-
Oracle
-
Other DatabaseUse Vanna to generate queries for any SQL database
vn.connect_to_snowflake(
account="myaccount",
username="myusername",
password="mypassword",
database="mydatabase",
role="myrole",
)
Training
You only need to train once. Do not train again unless you want to add more training data.
# The information schema query may need some tweaking depending on your database. This is a good starting point.
df_information_schema = vn.run_sql("SELECT * FROM INFORMATION_SCHEMA.COLUMNS")
# This will break up the information schema into bite-sized chunks that can be referenced by the LLM
plan = vn.get_training_plan_generic(df_information_schema)
plan
# If you like the plan, then uncomment this and run it to train
# vn.train(plan=plan)
# The following are methods for adding training data. Make sure you modify the examples to match your database.
# DDL statements are powerful because they specify table names, colume names, types, and potentially relationships
vn.train(ddl="""
CREATE TABLE IF NOT EXISTS my-table (
id INT PRIMARY KEY,
name VARCHAR(100),
age INT
)
""")
# Sometimes you may want to add documentation about your business terminology or definitions.
vn.train(documentation="Our business defines OTIF score as the percentage of orders that are delivered on time and in full")
# You can also add SQL queries to your training data. This is useful if you have some queries already laying around. You can just copy and paste those from your editor to begin generating new SQL.
vn.train(sql="SELECT * FROM my-table WHERE name = 'John Doe'")
# At any time you can inspect what training data the package is able to reference
training_data = vn.get_training_data()
training_data
# You can remove training data if there's obsolete/incorrect information.
vn.remove_training_data(id='1-ddl')
```## Asking the AI
Whenever you ask a new question, it will find the 10 most relevant pieces of training data and use it as part of the LLM prompt to generate the SQL.
```python
vn.ask(question=...)
Launch the User Interface
from vanna.flask import VannaFlaskApp
app = VannaFlaskApp(vn)
app.run()