@@ -28,26 +28,58 @@ On the other hand, knowledge graphs contain carefully structured data and are
2828designed to capture intricate relationships among discrete and seemingly
2929unrelated information.
3030
31- ArangoDB 's unique capabilities and flexible integration of knowledge graphs and
31+ Arango 's unique capabilities and flexible integration of knowledge graphs and
3232LLMs provide a powerful and efficient solution for anyone seeking to extract
3333valuable insights from diverse datasets.
3434
35- The GraphRAG component of AI Suite brings all the capabilities
35+ The GraphRAG component of the AI Suite brings all the capabilities
3636together with an easy-to-use interface, so you can make the knowledge accessible
3737to your organization.
3838
39+ ## Why GraphRAG
40+
3941GraphRAG is particularly valuable for use cases like the following:
42+
4043- Applications requiring in-depth knowledge retrieval
4144- Contextual question answering
4245- Reasoning over interconnected information
46+ - Discovery of relationships between concepts across documents
47+
48+ For detailed business scenarios, see [ GraphRAG Use Cases] ( use-cases.md ) .
49+
50+ ## Ways to use GraphRAG
51+
52+ You can interact with Arango's GraphRAG solution via a web interface or an API,
53+ depending on your needs.
54+
55+ ### Web Interface
56+
57+ The [ Web Interface] ( web-interface.md ) provides a user-friendly, no-code way to work
58+ with GraphRAG.
59+
60+ The web interface guides you through the process of the following:
61+
62+ 1 . Creating projects.
63+ 2 . Configuring Importer and Retriever services.
64+ 3 . Uploading documents to build knowledge graphs.
65+ 4 . Querying your knowledge graph with natural language.
66+ 5 . Exploring the graph structure visually.
67+
68+ ### API and Services
69+
70+ The [ AI Orchestrator] ( ../reference/ai-orchestrator.md ) ,
71+ [ Importer] ( ../reference/importer.md ) , and [ Retriever] ( ../reference/retriever.md )
72+ services provide programmatic access to create and manage GraphRAG pipelines,
73+ and give you access to advanced search methods.
4374
4475## How GraphRAG works
4576
46- ArangoDB 's GraphRAG solution democratizes the creation and usage of knowledge
77+ Arango 's GraphRAG solution democratizes the creation and usage of knowledge
4778graphs with a unique combination of vector search, graphs, and LLMs (privately or publicly hosted)
4879in a single product.
4980
5081The overall workflow involves the following steps:
82+
51831 . ** Chunking** :
5284 - Breaking down raw documents into text chunks
53852 . ** Entity and relation extraction for Knowledge Graph construction** :
@@ -76,82 +108,57 @@ information in a structured graph format, allowing efficient querying and retrie
76108For detailed information about the service, see the
77109[ Importer] ( ../reference/importer.md ) service documentation.
78110
79- ### Extract information from the Knowledge Graph
80-
81- The Retriever service enables intelligent search and retrieval of information
82- from your previously created Knowledge Graph.
83- You can extract information from Knowledge Graphs using two distinct methods:
84- - Global retrieval
85- - Local retrieval
86-
87- For detailed information about the service, see the
88- [ Retriever] ( ../reference/retriever.md ) service documentation.
89-
90- #### Global retrieval
91-
92- Global retrieval focuses on:
93- - Extracting information from the entire Knowledge Graph, regardless of specific
94- contexts or constraints.
95- - Provides a comprehensive overview and answers queries that span across multiple
96- entities and relationships in the graph.
97-
98- ** Use cases:**
99- - Answering broad questions that require a holistic understanding of the Knowledge Graph.
100- - Aggregating information from diverse parts of the Knowledge Graph for high-level insights.
101-
102- ** Example query:**
103-
104- Global retrieval can answer questions like _ ** What are the main themes or topics covered in the document** _ ?
111+ ### Query your Knowledge Graph
105112
106- During import, the entire Knowledge Graph is analyzed to identify and summarize
107- the dominant entities, their relationships, and associated themes. Global
108- retrieval uses these community summaries to answer questions from different
109- perspectives, then the information gets aggregated into the final response.
113+ The Retriever service enables intelligent search and retrieval using multiple
114+ search methods optimized for different query types. For detailed information
115+ about the service, see the [ Retriever] ( ../reference/retriever.md ) service documentation.
110116
111- #### Local retrieval
117+ The Retriever provides different search methods, each optimized for specific query patterns:
112118
113- Local retrieval is a more focused approach for:
114- - Queries that are constrained to specific subgraphs or contextual clusters
115- within the Knowledge Graph.
116- - Targeted and precise information extraction, often using localized sections
117- of the Knowledge Graph.
119+ - ** Instant Search** : Fast streaming responses for quick answers.
120+ - ** Deep Search** : LLM-orchestrated multi-step research for comprehensive accuracy.
121+ - ** Global Search** : Community-based analysis for themes and overviews.
122+ - ** Local Search** : Entity-focused retrieval for specific relationships.
118123
119- ** Use cases:**
120- - Answering detailed questions about a specific entity or a related group of entities.
121- - Retrieving information relevant to a particular topic or section in the Knowledge Graph.
124+ {{< info >}}
125+ The Web Interface exposes ** Instant Search** and ** Deep Search** as the primary
126+ methods for ease of use. For access to all search methods with advanced
127+ parameters, use the API directly. See [ Retriever - Search Methods] ( ../reference/retriever.md#search-methods )
128+ for complete details.
129+ {{< /info >}}
122130
123- ** Example query: **
131+ ## LLM Options
124132
125- Local retrieval can answer questions like _ ** What is the relationship between entity X and entity Y** _ ?
133+ The GraphRAG services can utilize public and private LLMs, depending on your
134+ infrastructure requirements and data governance needs.
126135
127- Local queries use hybrid search (semantic and lexical) over the Entities
128- collection, and then it expands that subgraph over related entities, relations
129- (and its LLM-generated verbal descriptions), text chunks, and communities.
136+ ### Self-hosted models via Triton Inference Server
130137
131- ### Private LLMs
138+ For air-gapped environments or strict data privacy requirements, you can run
139+ all models on your own infrastructure.
140+ The Triton Inference Server serves as the backbone for running your LLM
141+ and embedding models on your own machines. It handles all model operations, from
142+ processing text to generating embeddings, and provides both HTTP and gRPC interfaces
143+ for communication.
132144
133- If you're working in an air-gapped environment or need to keep your data
134- private, you can use the private LLM mode with
135- [ Triton Inference Server] ( ../reference/triton-inference-server.md ) .
145+ For setup instructions, see [ Triton Inference Server] ( ../reference/triton-inference-server.md )
146+ and [ MLflow] ( ../reference/mlflow.md ) documentation.
136147
137- This option allows you to run the service completely within your own
138- infrastructure. The Triton Inference Server is a crucial component when
139- running in private LLM mode. It serves as the backbone for running your
140- language (LLM) and embedding models on your own machines, ensuring your
141- data never leaves your infrastructure. The server handles all the complex
142- model operations, from processing text to generating embeddings, and provides
143- both HTTP and gRPC interfaces for communication.
148+ ### Using OpenAI-compatible endpoints
144149
145- ### Public LLMs
150+ For a simpler setup, you can use any service that exposes an OpenAI-compatible API
151+ endpoint. This includes both cloud providers and private corporate LLMs such as
152+ OpenAI, OpenRouter, Google Gemini, Anthropic Claude, and any corporate or self-hosted
153+ LLM with OpenAI-compatible endpoints.
146154
147- Alternatively, if you prefer a simpler setup and don't have specific privacy
148- requirements, you can use the public LLM mode. This option connects to cloud-based
149- services like OpenAI's models via the OpenAI API or a large array of models
150- (Gemini, Anthropic, publicly hosted open-source models, etc.) via the OpenRouter option.
155+ For detailed configuration examples, see:
156+ - [ Importer - Deployment Options] ( ../reference/importer.md#deployment-options )
157+ - [ Retriever - Installation] ( ../reference/retriever.md#installation )
151158
152159## Limitations
153160
154- The pre-release version of ArangoDB GraphRAG has the following limitations:
161+ The pre-release version of Arango GraphRAG has the following limitations:
155162
156163- You can only import a single file.
157164- The knowledge graph generated from the file is imported into a named graph
0 commit comments