Skip to content

Commit 8dd2eff

Browse files
nerpaulaSimran-B
andauthored
AI Suite | GraphRAG : Update and fix terminology, descriptions, examples, parameters, tutorial (#838)
* update terminology, descriptions, examples, parameters, tutorial * Review * LLM options in UI --------- Co-authored-by: Simran Spiller <simran@arangodb.com>
1 parent 04f4699 commit 8dd2eff

File tree

6 files changed

+208
-111
lines changed

6 files changed

+208
-111
lines changed

site/content/ai-suite/graphrag/technical-overview.md

Lines changed: 71 additions & 64 deletions
Original file line numberDiff line numberDiff line change
@@ -28,26 +28,58 @@ On the other hand, knowledge graphs contain carefully structured data and are
2828
designed to capture intricate relationships among discrete and seemingly
2929
unrelated information.
3030

31-
ArangoDB's unique capabilities and flexible integration of knowledge graphs and
31+
Arango's unique capabilities and flexible integration of knowledge graphs and
3232
LLMs provide a powerful and efficient solution for anyone seeking to extract
3333
valuable insights from diverse datasets.
3434

35-
The GraphRAG component of AI Suite brings all the capabilities
35+
The GraphRAG component of the AI Suite brings all the capabilities
3636
together with an easy-to-use interface, so you can make the knowledge accessible
3737
to your organization.
3838

39+
## Why GraphRAG
40+
3941
GraphRAG is particularly valuable for use cases like the following:
42+
4043
- Applications requiring in-depth knowledge retrieval
4144
- Contextual question answering
4245
- Reasoning over interconnected information
46+
- Discovery of relationships between concepts across documents
47+
48+
For detailed business scenarios, see [GraphRAG Use Cases](use-cases.md).
49+
50+
## Ways to use GraphRAG
51+
52+
You can interact with Arango's GraphRAG solution via a web interface or an API,
53+
depending on your needs.
54+
55+
### Web Interface
56+
57+
The [Web Interface](web-interface.md) provides a user-friendly, no-code way to work
58+
with GraphRAG.
59+
60+
The web interface guides you through the process of the following:
61+
62+
1. Creating projects.
63+
2. Configuring Importer and Retriever services.
64+
3. Uploading documents to build knowledge graphs.
65+
4. Querying your knowledge graph with natural language.
66+
5. Exploring the graph structure visually.
67+
68+
### API and Services
69+
70+
The [AI Orchestrator](../reference/ai-orchestrator.md),
71+
[Importer](../reference/importer.md), and [Retriever](../reference/retriever.md)
72+
services provide programmatic access to create and manage GraphRAG pipelines,
73+
and give you access to advanced search methods.
4374

4475
## How GraphRAG works
4576

46-
ArangoDB's GraphRAG solution democratizes the creation and usage of knowledge
77+
Arango's GraphRAG solution democratizes the creation and usage of knowledge
4778
graphs with a unique combination of vector search, graphs, and LLMs (privately or publicly hosted)
4879
in a single product.
4980

5081
The overall workflow involves the following steps:
82+
5183
1. **Chunking**:
5284
- Breaking down raw documents into text chunks
5385
2. **Entity and relation extraction for Knowledge Graph construction**:
@@ -76,82 +108,57 @@ information in a structured graph format, allowing efficient querying and retrie
76108
For detailed information about the service, see the
77109
[Importer](../reference/importer.md) service documentation.
78110

79-
### Extract information from the Knowledge Graph
80-
81-
The Retriever service enables intelligent search and retrieval of information
82-
from your previously created Knowledge Graph.
83-
You can extract information from Knowledge Graphs using two distinct methods:
84-
- Global retrieval
85-
- Local retrieval
86-
87-
For detailed information about the service, see the
88-
[Retriever](../reference/retriever.md) service documentation.
89-
90-
#### Global retrieval
91-
92-
Global retrieval focuses on:
93-
- Extracting information from the entire Knowledge Graph, regardless of specific
94-
contexts or constraints.
95-
- Provides a comprehensive overview and answers queries that span across multiple
96-
entities and relationships in the graph.
97-
98-
**Use cases:**
99-
- Answering broad questions that require a holistic understanding of the Knowledge Graph.
100-
- Aggregating information from diverse parts of the Knowledge Graph for high-level insights.
101-
102-
**Example query:**
103-
104-
Global retrieval can answer questions like _**What are the main themes or topics covered in the document**_?
111+
### Query your Knowledge Graph
105112

106-
During import, the entire Knowledge Graph is analyzed to identify and summarize
107-
the dominant entities, their relationships, and associated themes. Global
108-
retrieval uses these community summaries to answer questions from different
109-
perspectives, then the information gets aggregated into the final response.
113+
The Retriever service enables intelligent search and retrieval using multiple
114+
search methods optimized for different query types. For detailed information
115+
about the service, see the [Retriever](../reference/retriever.md) service documentation.
110116

111-
#### Local retrieval
117+
The Retriever provides different search methods, each optimized for specific query patterns:
112118

113-
Local retrieval is a more focused approach for:
114-
- Queries that are constrained to specific subgraphs or contextual clusters
115-
within the Knowledge Graph.
116-
- Targeted and precise information extraction, often using localized sections
117-
of the Knowledge Graph.
119+
- **Instant Search**: Fast streaming responses for quick answers.
120+
- **Deep Search**: LLM-orchestrated multi-step research for comprehensive accuracy.
121+
- **Global Search**: Community-based analysis for themes and overviews.
122+
- **Local Search**: Entity-focused retrieval for specific relationships.
118123

119-
**Use cases:**
120-
- Answering detailed questions about a specific entity or a related group of entities.
121-
- Retrieving information relevant to a particular topic or section in the Knowledge Graph.
124+
{{< info >}}
125+
The Web Interface exposes **Instant Search** and **Deep Search** as the primary
126+
methods for ease of use. For access to all search methods with advanced
127+
parameters, use the API directly. See [Retriever - Search Methods](../reference/retriever.md#search-methods)
128+
for complete details.
129+
{{< /info >}}
122130

123-
**Example query:**
131+
## LLM Options
124132

125-
Local retrieval can answer questions like _**What is the relationship between entity X and entity Y**_?
133+
The GraphRAG services can utilize public and private LLMs, depending on your
134+
infrastructure requirements and data governance needs.
126135

127-
Local queries use hybrid search (semantic and lexical) over the Entities
128-
collection, and then it expands that subgraph over related entities, relations
129-
(and its LLM-generated verbal descriptions), text chunks, and communities.
136+
### Self-hosted models via Triton Inference Server
130137

131-
### Private LLMs
138+
For air-gapped environments or strict data privacy requirements, you can run
139+
all models on your own infrastructure.
140+
The Triton Inference Server serves as the backbone for running your LLM
141+
and embedding models on your own machines. It handles all model operations, from
142+
processing text to generating embeddings, and provides both HTTP and gRPC interfaces
143+
for communication.
132144

133-
If you're working in an air-gapped environment or need to keep your data
134-
private, you can use the private LLM mode with
135-
[Triton Inference Server](../reference/triton-inference-server.md).
145+
For setup instructions, see [Triton Inference Server](../reference/triton-inference-server.md)
146+
and [MLflow](../reference/mlflow.md) documentation.
136147

137-
This option allows you to run the service completely within your own
138-
infrastructure. The Triton Inference Server is a crucial component when
139-
running in private LLM mode. It serves as the backbone for running your
140-
language (LLM) and embedding models on your own machines, ensuring your
141-
data never leaves your infrastructure. The server handles all the complex
142-
model operations, from processing text to generating embeddings, and provides
143-
both HTTP and gRPC interfaces for communication.
148+
### Using OpenAI-compatible endpoints
144149

145-
### Public LLMs
150+
For a simpler setup, you can use any service that exposes an OpenAI-compatible API
151+
endpoint. This includes both cloud providers and private corporate LLMs such as
152+
OpenAI, OpenRouter, Google Gemini, Anthropic Claude, and any corporate or self-hosted
153+
LLM with OpenAI-compatible endpoints.
146154

147-
Alternatively, if you prefer a simpler setup and don't have specific privacy
148-
requirements, you can use the public LLM mode. This option connects to cloud-based
149-
services like OpenAI's models via the OpenAI API or a large array of models
150-
(Gemini, Anthropic, publicly hosted open-source models, etc.) via the OpenRouter option.
155+
For detailed configuration examples, see:
156+
- [Importer - Deployment Options](../reference/importer.md#deployment-options)
157+
- [Retriever - Installation](../reference/retriever.md#installation)
151158

152159
## Limitations
153160

154-
The pre-release version of ArangoDB GraphRAG has the following limitations:
161+
The pre-release version of Arango GraphRAG has the following limitations:
155162

156163
- You can only import a single file.
157164
- The knowledge graph generated from the file is imported into a named graph

0 commit comments

Comments
 (0)