What is GraphRAG ? Everything You Need To Know
Large Language Models (LLMs) often "hallucinate," creating convincing but inaccurate answers when they lack up-to-date information. RAG (Retrieval-Augmented Generation) helps reduce this by connecting LLMs to external data sources, pulling in factual knowledge to improve accuracy. Traditional RAG, however, retrieves isolated text chunks, which can miss the bigger picture and relationships between data.
Graph RAG enhances this by using graph-based retrieval. It pulls in connected information, like nodes and paths from a graph structure, capturing relationships to generate more accurate, context-rich responses. This approach also keeps input data concise by summarizing it through the graph’s structure.
What is GraphRAG?
GraphRAG takes RAG to the next level. Instead of just pulling in chunks of text like traditional RAG, GraphRAG dives into a web of relationships—a knowledge graph where everything’s connected. So when you ask a question, GraphRAG doesn’t just find the answer; it understands how different pieces of information are linked to give you a smarter, more complete answer.
Picture this: you’re trying to summarize why a big event, like a company merger, went down the way it did. Traditional RAG might dig up reports about each company, the merger process, and some outcomes. But GraphRAG? It’s way more clever. It doesn’t just throw facts at you—it connects the dots. It shows you how the leadership styles of the companies meshed, how market conditions nudged things along, and how the customer bases blended together. Instead of a jumble of facts, you get a clear, connected story that really explains what happened.
And here’s the cool part: GraphRAG doesn’t just give you a pile of text—it hands you the full picture, neatly connected. When you need to summarize something complex, like why a merger succeeded, GraphRAG pulls in everything relevant and shows you how it all fits together. It’s like getting the whole story in one shot, with all the juicy connections that make it make sense.
What Are The Types of GraphRAG
There are different types of GraphRAG based on different domains and applications, we summarize as follows:
Query-based GraphRAG: This approach starts by transforming your natural language question into a graph-based query such as Gremlin or Cypher. Then graph engine runs this query against the graph. Finally, it feeds these results into the LLM, which uses them to generate a final answer.
Content-based GraphRAG: Instead of just running a query, this type focuses on retrieving different layers of graph content—things like individual nodes, relationships (triplets), paths through the graph, or even some subgraphs. These various elements of the graph are then fed into the LLM, allowing it to consider multiple angles and nuances to generate a more informed and comprehensive answer.
Agentic GraphRAG: This is where things get really interesting. In Agentic GraphRAG, the graph retrieval process becomes a tool in the hands of an LLM-based AI agent. The agent doesn’t just passively use the graph information—it actively plans and reasons with it, adapting its approach to come up with the best possible answer. With tools like PuppyGraph’s open-source library, PuppyGraphAgent, setting up this kind of dynamic and intelligent system becomes much more accessible.
This variety in GraphRAG approaches means it can be tailored to fit the specific needs and complexities of different tasks, providing flexibility and depth that go beyond traditional RAG methods.
How Does GraphRAG Work ?
Prerequisites
Follow the guide https://docs.puppygraph.com/getting-started/querying-postgresql-data-as-a-graph#deployment to install both PostgreSQL and PuppyGraph.
We will prepare following necessary softwares:
- PostgreSQL: Serves our relational database to hold the actual data. You can also choose any of your favourite database
- PuppyGraph: Enables Graph Query on top the relationship database. The previous instruction starts puppygraph as well. If you would like to run it separately, follow the guide here https://docs.puppygraph.com/getting-started/launching-puppygraph-in-docker to install.
Python Libraries: Enables the actual GraphRAG
# Contains puppygraph client and puppygraph agent
pip3 install puppygraph
# Langchain version of OpenAI wrapper
pip3 install langchain-openai
OpenAI Access Token: Make sure OPENAI_API_KEY
is set as an environment variable.
Load Dataset into PostgreSQL
We will store the imdb dataset into the PostgreSQL first.
Download the dataset
Download the dataset from IMDB with these two links and unarchive them. This will download around 1GB data from the Internet.
wget https://datasets.imdbws.com/name.basics.tsv.gz
wget https://datasets.imdbws.com/title.principals.tsv.gz
wget https://datasets.imdbws.com/title.basics.tsv.gz
gzip -d name.basics.tsv.gz
gzip -d title.principals.tsv.gz
gzip -d title.basics.tsv.gz
docker cp name.basics.tsv postgres:/name.basics.tsv
docker cp title.principals.tsv postgres:/title.principals.tsv
docker cp title.basics.tsv postgres:/title.basics.tsv
Load the dataset into PostgreSQL
Once we have the dataset, we first define the tables in PostgreSQL where the data will be stored. Start a Postgres shell by running the following command, input the password (by default postgres123
) and then create tables in the database.
docker exec -it postgres psql -h postgres -U postgres
SQL for data preparation in PostgreSQL
CREATE TABLE name_basics (
nconst TEXT PRIMARY KEY,
primaryName TEXT,
birthYear INTEGER,
deathYear INTEGER,
primaryProfession TEXT[], -- Array to store CSV values
knownForTitles TEXT[] -- Array to store CSV values
);
CREATE TEMP TABLE tmp_name_basics (
nconst TEXT,
primaryName TEXT,
birthYear INTEGER,
deathYear INTEGER,
primaryProfession TEXT, -- Temporarily store as TEXT
knownForTitles TEXT -- Temporarily store as TEXT
);
\copy tmp_name_basics FROM '/name.basics.tsv' WITH (FORMAT csv, DELIMITER E'\t', HEADER true, NULL '\N', QUOTE E'\b');
INSERT INTO name_basics (nconst, primaryName, birthYear, deathYear, primaryProfession, knownForTitles)
SELECT
nconst,
primaryName,
birthYear,
deathYear,
string_to_array(primaryProfession, ',')::TEXT[], -- Convert CSV text to array
string_to_array(knownForTitles, ',')::TEXT[] -- Convert CSV text to array
FROM
tmp_name_basics;
CREATE TABLE title_basics (
tconst TEXT PRIMARY KEY,
titleType TEXT,
primaryTitle TEXT,
originalTitle TEXT,
isAdult BOOLEAN,
startYear INTEGER,
endYear INTEGER,
runtimeMinutes INTEGER,
genres TEXT[] -- Array to store CSV values
);
CREATE TEMP TABLE tmp_title_basics (
tconst TEXT PRIMARY KEY,
titleType TEXT,
primaryTitle TEXT,
originalTitle TEXT,
isAdult TEXT,
startYear INTEGER,
endYear INTEGER,
runtimeMinutes INTEGER,
genres TEXT
);
\COPY tmp_title_basics FROM '/title.basics.tsv' WITH (FORMAT csv, DELIMITER E'\t', HEADER true, NULL '\N', QUOTE E'\b');
INSERT INTO title_basics (tconst, titleType, primaryTitle, originalTitle, isAdult, startYear, endYear, runtimeMinutes, genres)
SELECT
tconst,
titleType,
primaryTitle,
originalTitle,
CASE isAdult WHEN '0' THEN FALSE ELSE TRUE END AS isAdult, -- Convert text to boolean
startYear,
endYear,
runtimeMinutes,
string_to_array(genres, ',') AS genres -- Convert CSV text to array
FROM tmp_title_basics;
CREATE TABLE title_principals (
tconst TEXT,
ordering INTEGER,
nconst TEXT,
category TEXT,
job TEXT,
characters TEXT,
primary key (tconst, ordering)
);
\COPY title_principals FROM '/title.principals.tsv' WITH (FORMAT csv, DELIMITER E'\t', HEADER true, NULL '\N',QUOTE E'\b');
The process loads raw TSV files into temporary tables and then perform some data transformation and cleanup to get the data tables.
Converting PostgreSQL tables to a Graph with PuppyGraph
We used the following schema to create the graph within PuppyGraph. The schema contains 2 different types of nodes: failure_type
, asset
, and work_order
. As well as 1 type of relationship: cast_and_crew
. Each of the nodes and relationships is powered by one or more tables in PostgreSQL. PuppyGraph provides a graphical schema builder for users to define the mapping.
- Open PuppyGraph web (http://localhost:8081), in the schema tab, use the Upload Graph Schema JSON
- Upload the JSON schema, make sure to use the correct user name, password and jdbc uri for your PostgreSQL.
{
"catalogs": [{ "name": "imdb", "type": "postgresql", "jdbc": { ... } }],
"graph": {
"vertices": [{ "label": "person", "oneToOne": { ... } }, { "label": "title", "oneToOne": { ... } }],
"edges": [{ "label": "cast_and_crew", "fromVertex": "title", "toVertex": "person", "tableSource": { ... } }]
}
}
This is a highly abbreviated version of the schema, retaining only the most critical elements. For the full schema, please refer to the following:
{
"catalogs": [
{
"name": "imdb",
"type": "postgresql",
"jdbc": {
"username": "postgres",
"password": "postgres123",
"jdbcUri": "jdbc:postgresql://postgres:5432/postgres",
"driverClass": "org.postgresql.Driver"
}
}
],
"graph": {
"vertices": [
{
"label": "person",
"oneToOne": {
"tableSource": {
"catalog": "imdb",
"schema": "public",
"table": "name_basics"
},
"id": {
"fields": [
{
"type": "String",
"field": "nconst",
"alias": "nconst"
}
]
},
"attributes": [
{
"type": "String",
"field": "primaryName",
"alias": "primaryName"
},
{
"type": "Int",
"field": "birthYear",
"alias": "birthYear"
},
{
"type": "Int",
"field": "deathYear",
"alias": "deathYear"
}
]
}
},
{
"label": "title",
"oneToOne": {
"tableSource": {
"catalog": "imdb",
"schema": "public",
"table": "title_basics"
},
"id": {
"fields": [
{
"type": "String",
"field": "tconst",
"alias": "tconst"
}
]
},
"attributes": [
{
"type": "String",
"field": "titleType",
"alias": "titleType"
},
{
"type": "String",
"field": "primaryTitle",
"alias": "primaryTitle"
},
{
"type": "String",
"field": "originalTitle",
"alias": "originalTitle"
},
{
"type": "Boolean",
"field": "isAdult",
"alias": "isAdult"
},
{
"type": "Int",
"field": "startYear",
"alias": "startYear"
},
{
"type": "Int",
"field": "endYear",
"alias": "endYear"
},
{
"type": "Int",
"field": "runtimeMinutes",
"alias": "runtimeMinutes"
}
]
}
}
],
"edges": [
{
"label": "cast_and_crew",
"fromVertex": "title",
"toVertex": "person",
"tableSource": {
"catalog": "imdb",
"schema": "public",
"table": "title_principals"
},
"id": {
"fields": [
{
"alias": "id_title",
"field": "tconst",
"type": "String"
},
{
"alias": "id_order",
"field": "ordering",
"type": "Int"
}
]
},
"fromId":{
"fields": [
{
"type": "String",
"field": "tconst",
"alias": "from"
}
]
},
"toId":{
"fields": [
{
"type": "String",
"field": "nconst",
"alias": "to"
}
]
},
"attributes": [
{
"type": "Int",
"field": "ordering",
"alias": "ordering"
},
{
"type": "String",
"field": "category",
"alias": "category"
},
{
"type": "String",
"field": "job",
"alias": "job"
},
{
"type": "String",
"field": "characters",
"alias": "characters"
}
]
}
]
}
}
PuppyGraph shows the schema visualization as follows:
Also in the dashboard tab, we could see the edge and node examples as follows:
Now that our graph creation is finished. We will go to the actual meaty part to set up the GraphRAG agent using Python.
Setup the PuppyGraphClient
Setup the client with necessary credentials
from puppygraph import PuppyGraphClient, PuppyGraphHostConfig
client = PuppyGraphClient(
PuppyGraphHostConfig(
ip="localhost", username="puppygraph", password="puppygraph123"
)
)
Querying the Graph
Then, we are able to query the graph using either Cypher or Gremlin via the client.
>>> client.gremlin_query("g.E().limit(1).elementMap()")
[{<T.id: 1>: 'cast_and_crew[tt5350914_8]', <T.label: 4>: 'cast_and_crew', <Direction.OUT: 'OUT'>: {<T.id: 1>: 'title[tt5350914]', <T.label: 4>: 'title'}, <Direction.IN: 'IN'>: {<T.id: 1>: 'person[nm1916604]', <T.label: 4>: 'person'}, 'ordering': 8, 'category': 'actor', 'job': None, 'characters': None}]
>>> client.cypher_query("MATCH ()-[r]->() RETURN properties(r) LIMIT 1")
[{'properties(r)': {'category': 'actor', 'characters': '["Cosmo"]', 'ordering': 13}}]
Setup Query Tools
We setup tools that can be used by the GraphRAG agent
from langchain_core.tools import StructuredTool
from puppygraph import PuppyGraphClient
def _get_cypher_query_tool(puppy_graph_client: PuppyGraphClient):
"""Get the Cypher query tool."""
return StructuredTool.from_function(
func=puppy_graph_client.cypher_query,
name="query_graph_cypher",
description="Query the graph database using Cypher.",
args_schema=create_model(
"", query=(str, Field(description="The Cypher query to run"))
),
)
def _get_gremlin_query_tool(puppy_graph_client: PuppyGraphClient):
"""Get the Gremlin query tool."""
return StructuredTool.from_function(
func=puppy_graph_client.gremlin_query,
name="query_graph_gremlin",
description="Query the graph database using Gremlin.",
args_schema=create_model(
"", query=(str, Field(description="The Gremlin query to run"))
),
)
(Note: These two tools are for demonstration purposes only. In reality, you don't have to define these tools explicitly, we've already defined these for you within the PuppyGraphAgent.)
Setup LLM
We use openai's GPT4o-2024-08-06
model :
from langchain_openai import ChatOpenAI
llm = ChatOpenAI(
model="gpt-4o-2024-08-06",
temperature=0,
)
Setup Prompts
We need to do some prompt engineering for LLM to understand the graph.
Setup Graph Schema Prompt
We first setup the graph schema prompt with semantic information around the node and edges and attributes of the graph.
GRAPH_SCHEMA_PROMPT = """
Nodes are the following:
- person:
properties:
- name: primaryName
type: String
description: The name of the person, as listed in the IMDb database.
- name: birthYear
type: Int
description: The birth year of the person (if available).
- name: deathYear
type: Int
description: The death year of the person (if available).
- title:
properties:
- name: titleType
type: String
description: The type/format of the title (e.g., movie, short, tvseries, tvepisode, video, etc.).
- name: primaryTitle
type: String
description: The more popular title or the title used by filmmakers on promotional materials at the point of release.
- name: originalTitle
type: String
description: The original title, in the original language.
- name: isAdult
type: Boolean
description: Indicates whether the title is for adults (1: adult title, 0: non-adult title).
- name: startYear
type: Int
description: Represents the release year of a title. For TV Series, this is the series start year.
- name: endYear
type: Int
description: For TV Series, this is the series end year. '\\N' for all other title types.
- name: runtimeMinutes
type: Int
description: The primary runtime of the title, in minutes.
Edges are the following:
- cast_and_crew:
from: title
to: person
properties:
- name: ordering
type: Int
description: A unique identifier for the row, used to determine the order of people associated with this title.
- name: category
type: String
description: The category of job that the person was in (e.g., actor, director).
- name: job
type: String
description: The specific job title if applicable, else '\\N'.
- name: characters
type: String
description: The name of the character played if applicable, else '\\N'.
The relationships are the following:
g.V().hasLabel('title').out('cast_and_crew').hasLabel('person'),
g.V().hasLabel('person').in('cast_and_crew').hasLabel('title'),
If filter by category, you must use outE() or inE(), because the category is stored in the EDGE properties.
"""
Setup the agent prompt
We use a Chain-of-Thought technique to setup the agent prompt, asking the LLM to use think-plan-tool methodology to do the reasoning and generation.
from langchain_core.prompts.chat import ChatPromptTemplate, MessagesPlaceholder
chat_prompt_template = ChatPromptTemplate.from_messages(
[
(
"system",
"You are a helpful assistant to help answer user questions about imdb."
"You will need to use the information stored in the graph database to answer the user's questions."
"Here is some information about the graph database schema.\n"
f"{GRAPH_SCHEMA_PROMPT}",
),
(
"system",
"You must first output a PLAN, then you can use the PLAN to call the tools.\n"
"Each STEP of the PLAN should be corresponding to one or more function calls (but not less), either simple or complex.\n"
"Minimize the number of steps in the PLAN, but make sure the PLAN is workable.\n"
"Remember, each step can be converted to a Gremlin query, since Gremlin query can handle quite complex queries,"
"each step can be complex as well as long as it can be converted to a Gremlin query.",
),
MessagesPlaceholder(variable_name="message_history"),
(
"system",
"Always use the JSON format {\n"
"'THINKING': <the thought process in PLAIN TEXT>,"
"'PLAN': <the plan contains multiple steps in PLAIN TEXT, Your Original plan or Update plan after seeing some executed results>,"
f"'CONCLUSION': <Keep your conclusion simple and clear if you decide to conclude >",
),
],
template_format="jinja2",
)
Note that “message_history” is a variable that is created and maintained by PuppyGraphAgent, so we need to make sure it is a placeholder within the prompt.
Combine everything together with PuppyGraph Agent
Now we've got the client, llm and prompt template, we can use the PuppyGraphAgent to combine different pieces together into the GraphRAG agent:
from puppygraph.rag import PuppyGraphAgent
pg_agent = PuppyGraphAgent(
puppy_graph_client=client,
llm=llm,
chat_prompt_template=chat_prompt_template
)
Final Results
Now we can start by querying the agent and having a conversation with the agent
pg_agent.query("how many movies does Tom hanks acted in?")
pg_agent.query("how many of them are directed by Steven Spielberg?")
pg_agent.query("What's tom hanks recent movie in 2024?")
pg_agent.query("all movies in 2024 by him")
pg_agent.reset_messages()
We can integrate our agent with Gradio, and here are some interesting results:
We can see that PuppyGraphAgent not only come up with the final result, but also, plans, reasons and sometimes doing multiple graph queries or correct the queries if the results are undesired as shown in the intermediate result. It really shows the potential of Agentic GraphRAG, where it is more dynamic and intelligent compared to the standard GraphRAG.
Further prompt engineering might be necessary to have a more robust GraphRAG agent, including few-shot examples, in which we aim to release in our next version.
Advantages of GraphRAG
GraphRAG takes things up a notch compared to traditional RAG, offering several key advantages. Unlike traditional RAG, which pulls in isolated text, GraphRAG digs into the connections between pieces of information, giving you more accurate and meaningful answers. It excels at answering complex questions by understanding how different pieces of data relate, providing a more complete answer. Plus, GraphRAG focuses on the most relevant information, reducing the redundancy and overwhelming text that traditional RAG can produce. This ensures you get concise, targeted results without the extra noise.
Another major advantage is its flexibility in data integration. With PuppyGraph, you can easily map data from relational databases into a graph structure without needing complex ETL processes, making it simple to work with your existing data. Moreover, GraphRAG, especially with an agent framework like PuppyGraphAgent, doesn’t just retrieve information—it plans, adapts, and intelligently reasons through multiple queries to provide the best possible answers. In short, GraphRAG offers smarter, more connected answers, making it a powerful upgrade over traditional RAG.
Challenges in GraphRAG
GraphRAG is a game-changer, but it doesn’t come without its challenges. To really make the most of it, there are some significant hurdles to clear.
First, there’s the issue of complex knowledge graph creation. Building large-scale knowledge graphs from a mix of different data sources is no small feat. Typically, this requires extensive ETL (Extract, Transform, Load) processes, which can be time-consuming and resource-heavy. This complexity can slow down getting your GraphRAG up and running, making it harder to manage and update the graph data.
Next, scalability and latency issues become a real concern as your data grows to petabyte scale. Ensuring quick query responses while dealing with massive datasets, especially when you’re running complex, multi-hop queries, is tough. Many systems struggle with this, making it a key issue for any serious GraphRAG implementation.
Then, there’s the challenge of building intelligent, adaptive reasoning systems. As your graph expands, the number of potential subgraphs can skyrocket, making it harder to zero in on the most relevant ones without overwhelming the system. Plus, accurately matching a textual query to the right graph data isn’t easy. It requires advanced algorithms that can grasp both the text and the complex structure of the data, all while reasoning and replanning on the fly. Developing a system that can handle all this requires sophisticated, flexible frameworks.
These challenges are significant, but overcoming them is key to unlocking the full potential of GraphRAG in any application.
How PuppyGraph can help with GraphRAG
PuppyGraph brings some serious firepower to the table when it comes to GraphRAG. Here’s how:
Zero ETL for Hassle-Free Graph Creation: PuppyGraph connects directly to all your data sources without the need for complex ETL (Extract, Transform, Load) processes. It treats all your data as a unified graph without needing to migrate anything into a dedicated graph database. This makes it super easy to create large-scale knowledge graphs and run queries, getting your GraphRAG up and running in as little as 10 minutes.
Agentic GraphRAG: Traditional RAG just retrieves and generates, but PuppyGraph’s Agentic GraphRAG takes it further. It’s goal-driven, meaning it plans, runs multiple graph queries, and reasons intelligently based on the results. If needed, it can even re-plan or summarize to get the best possible answer. Plus, there is our open-source library, PuppyGraphAgent, to help you set this up.
Built-In Support for Gremlin & Cypher: With PuppyGraph, you’re not locked into one query language. It supports both Gremlin and Cypher, so your GraphRAG application can pick the best tool for the job—or use both! This flexibility boosts accuracy and tailors queries to fit the specific input, making sure you get the most desired results.
Petabyte Scale with Ultra-Low Latency: PuppyGraph is built to handle massive amounts of data—think petabyte scale—with lightning speed. Even complex queries, like a 10-hop neighbor search, take just 2.26 seconds. This means you can dig out even the most hidden insights from your vast datasets, kind of like finding a needle in a haystack but at super speed.
Conclusion
Looking to the future, GraphRAG is poised to change the way we work with LLMs. As we continue to refine these methods, we’re on the brink of seeing even more powerful, connected, and insightful AI Agents. The future of AI is all about leveraging those connections, and GraphRAG is leading the charge toward a smarter, more integrated world of possibilities.
Get started with PuppyGraph!
Developer Edition
- Forever free
- Single noded
- Designed for proving your ideas
- Available via Docker install
Enterprise Edition
- 30-day free trial with full features
- Everything in developer edition & enterprise features
- Designed for production
- Available via AWS AMI & Docker install