Search...
ctrl/
Light
Dark
System
Sign in
Environment:

How to set up Gel AI in Python

Gel AI brings vector search capabilities and retrieval-augmented generation directly into the database. It's integrated into the Gel Python binding via the gel.ai module.

Copy
$ 
pip install 'gel[ai]'

AI is an Gel extension. To enable it, we will need to add the extension to the app’s schema:

Copy
using extension ai;

Gel AI uses external APIs in order to get vectors and LLM completions. For it to work, we need to configure an API provider and specify their API key. Let's open EdgeQL REPL and run the following query:

Copy
configure current database
insert ext::ai::OpenAIProviderConfig {
  secret := 'sk-....',
};

Now our Gel application can take advantage of OpenAI's API to implement AI capabilities.

Gel AI comes with its own UI that can be used to configure providers, set up prompts and test them in a sandbox.

Most API providers require you to set up and account and charge money for model use.

Before we start introducing AI capabilities, let's set up our database with a schema and populate it with some data (we're going to be helping Komi-san keep track of her friends).

Copy
module default {
    type Friend {
        required name: str {
            constraint exclusive;
        };

        summary: str;               # A brief description of personality and role
        relationship_to_komi: str;  # Relationship with Komi
        defining_trait: str;        # Primary character trait or quirk
    }
}

Here's a shell command you can paste and run that will populate the database with some sample data.

Copy
$ 
cat << 'EOF' > populate_db.edgeql
insert Friend {
    name := 'Tadano Hitohito',
    summary := 'An extremely average high school boy with a remarkable ability to read the atmosphere and understand others\' feelings, especially Komi\'s.',
    relationship_to_komi := 'First friend and love interest',
    defining_trait := 'Perceptiveness',
};

insert Friend {
    name := 'Osana Najimi',
    summary := 'An extremely outgoing person who claims to have been everyone\'s childhood friend. Gender: Najimi.',
    relationship_to_komi := 'Second friend and social catalyst',
    defining_trait := 'Universal childhood friend',
};

insert Friend {
    name := 'Yamai Ren',
    summary := 'An intense and sometimes obsessive classmate who is completely infatuated with Komi.',
    relationship_to_komi := 'Self-proclaimed guardian and admirer',
    defining_trait := 'Obsessive devotion',
};

insert Friend {
    name := 'Katai Makoto',
    summary := 'A intimidating-looking but shy student who shares many communication problems with Komi.',
    relationship_to_komi := 'Fellow communication-challenged friend',
    defining_trait := 'Scary appearance but gentle nature',
};

insert Friend {
    name := 'Nakanaka Omoharu',
    summary := 'A self-proclaimed wielder of dark powers who acts like an anime character and is actually just a regular gaming enthusiast.',
    relationship_to_komi := 'Gaming buddy and chuunibyou friend',
    defining_trait := 'Chuunibyou tendencies',
};
EOF
Copy
$ 
gel query -f populate_db.edgeql
Show more

In order to get Gel to produce embedding vectors, we need to create a special deferred index on the type we would like to perform similarity search on. More specifically, we need to specify an EdgeQL expression that produces a string that we're going to create an embedding vector for. This is how we would set up an index if we wanted to perform similarity search on Friend.summary:

Copy
Show 7 hidden lines...
        relationship_to_komi: str;  # Relationship with Komi
        defining_trait: str;        # Primary character trait or quirk

        deferred index ext::ai::index(embedding_model := 'text-embedding-3-small')
            on (.summary);
    }
}

But actually, in our case it would be better if we could similarity search across all properties at the same time. We can define the index on a more complex expression - like a concatenation of string properties - like this:

Copy
Show 8 hidden lines...
        defining_trait: str;        # Primary character trait or quirk

        deferred index ext::ai::index(embedding_model := 'text-embedding-3-small')
            on (.summary);
            on (
                .name ++ ' ' ++ .summary ++ ' '
                ++ .relationship_to_komi ++ ' '
                ++ .defining_trait
            );
    }
}

Once we're done with schema modification, we need to apply them by going through a migration:

Copy
$ 
gel migration create
Copy
$ 
gel migrate

That's it! Gel will make necessary API requests in the background and create an index that will enable us to perform efficient similarity search.

In order to run queries against the index we just created, we need to create a Gel client and pass it to a Gel AI instance.

Copy
import gel
import gel.ai

gel_client = gel.create_client()
gel_ai = gel.ai.create_rag_client(client)

text = "Who helps Komi make friends?"
vector = gel_ai.generate_embeddings(
    text,
    "text-embedding-3-small",
)

gel_client.query(
    "select ext::ai::search(Friend, <array<float32>>$embedding_vector",
    embedding_vector=vector,
)

We are going to execute a query that calls a single function: ext::ai::search(<type>, <search_vector>). That function accepts an embedding vector as the second argument, not a text string. This means that in order to similarity search for a string, we need to create a vector embedding for it using the same model as we used to create the index. The Gel AI binding in Python comes with a generate_embeddings function that does exactly that:

Copy
Show 3 hidden lines...
gel_client = gel.create_client()
gel_ai = gel.ai.create_rag_client(client)

text = "Who helps Komi make friends?"
vector = gel_ai.generate_embeddings(
    text,
    "text-embedding-3-small",
)

Now we can plug that vector directly into our query to get similarity search results:

Copy
Show 9 hidden lines...
    "text-embedding-3-small",
)

gel_client.query(
    "select ext::ai::search(Friend, <array<float32>>$embedding_vector",
    embedding_vector=vector,
)

One more feature Gel AI offers is built-in retrieval-augmented generation, also known as RAG.

Gel comes preconfigured to be able to process our text query, perform similarity search across the index we just created, pass the results to an LLM and return a response. In order to access the built-in RAG, we need to start by selecting an LLM and passing its name to the Gel AI instance constructor:

Copy
Show 3 hidden lines...
gel_client = gel.create_client()
gel_ai = gel.ai.create_rag_client(
    client,
    model="gpt-4-turbo-preview"
)

Now we can access the RAG using the query_rag function like this:

Copy
Show 6 hidden lines...
    model="gpt-4-turbo-preview"
)

gel_ai.query_rag(
    "Who helps Komi make friends?",
    context="Friend",
)

We can also stream the response like this:

Copy
Show 6 hidden lines...
    model="gpt-4-turbo-preview"
)

gel_ai.query_rag(
gel_ai.stream_rag(
    "Who helps Komi make friends?",
    context="Friend",
)

You are now sufficiently equipped to use Gel AI in your applications.

If you'd like to build something on your own, make sure to check out the Reference manual for the AI extension in order to learn the details about using different APIs and models, configuring prompts or using the UI. Make sure to take a look at the Python binding reference, too.

And if you would like more guidance for how Gel AI can be fit into an application, take a look at the FastAPI Gel AI Tutorial, where we're building a search bot using features you learned about above.