Zero-Maintenance AI Code Assistant - Local-first, cost-effective, privacy-safe
augment-lite-mcp is a zero-maintenance, local-first AI code assistant engine that integrates with AI programming tools like Claude Code via MCP (Model Context Protocol).
Zero-Maintenance Search + Local Privacy + Low-Cost AI = Ideal Programming Assistant
- π₯ Zero Maintenance: Auto-incremental indexing, no manual rebuild required
- π Privacy First: Code stored entirely locally (DuckDB + SQLite)
- π° Cost Effective: Local BM25+Vector first, LLM only for final filtering
- π― Hybrid Search: BM25 keywords + vector semantic dual matching
flowchart LR
subgraph Client["π₯οΈ Client"]
CC["Claude Code<br/>IDE"]
end
subgraph MCP["π‘ MCP Server"]
direction TB
API["31 MCP Tools"]
API --> RAG["RAG Engine"]
API --> Code["Code Analysis<br/>(Tree-sitter)"]
API --> Mem["Memory<br/>& Tasks"]
end
subgraph Index["π Index Layer"]
direction TB
BM25["BM25<br/>DuckDB FTS"]
Vec["Vector<br/>FAISS 2560d"]
Chunk["Chunks<br/>JSONL"]
end
subgraph LLM["π€ LLM Layer"]
direction TB
GLM["GLM-4.7<br/>(reason)"]
MM["MiniMax-M2.1<br/>(fast)"]
Emb["Qwen3-Embed<br/>(OpenRouter)"]
end
CC <-->|"MCP Protocol"| API
RAG --> BM25
RAG --> Vec
RAG --> GLM
RAG --> MM
Vec --> Emb
BM25 --> Chunk
Vec --> Chunk
style CC fill:#e3f2fd
style API fill:#fff3e0
style RAG fill:#e8f5e9
style Vec fill:#fce4ec
acemcp-inspired Zero-Maintenance Experience
# No manual rebuild needed, everything happens automatically
./scripts/manage.sh add auto . # Initial project add
# After any code modifications
# Changes are auto-detected and indexed on search- β Auto-detect file changes (mtime + MD5)
- β Only update changed files (60x faster)
- β Completely transparent, no user intervention
Local Vector + Remote LLM Smart Filtering
Layer 1: Vector Embeddings (OpenRouter API / Local fallback)
β BM25 + Vector hybrid search (kΓ3 over-fetch)
β ~50 candidates β per-file dedup β ~35 candidates
β Model: qwen/qwen3-embedding-4b (2560 dims, API)
β Fallback: all-MiniLM-L6-v2 (384 dims, local)
Layer 2: GLM-4.7 / MiniMax-M2.1 LLM Smart Filtering
β Semantic understanding + Re-rank
β Final Top-K high-quality results
β Uses OpenAI-compatible API format
Advantages:
- Local-first: BM25+Vector computed locally, no API costs
- LLM filtering: Only candidate results sent to LLM, reduced token consumption
- Optional fallback: Auto-fallback to pure BM25 when Vector not installed
Model Selection: Multiple embedding models supported, see Vector Models Comparison
flowchart TB
Q["π Query: auth module"]
Q --> BM25["<b>BM25</b><br/>DuckDB FTS<br/>Keyword matching"]
Q --> Vector["<b>Vector</b><br/>FAISS + Qwen3-Embedding<br/>Semantic similarity (2560 dims)"]
BM25 -->|"kΓ3 results"| Merge["<b>Hybrid Merge</b><br/>score = bm25_wΓs1 + vec_wΓs2<br/>(default 0.5 each)"]
Vector -->|"kΓ3 results"| Merge
Merge --> Dedup["<b>Per-file Dedup</b><br/>Max 2 chunks per file"]
Dedup --> Rerank["<b>LLM Re-rank</b><br/>GLM-4.7 / MiniMax"]
Rerank --> Result["π Return Top-K Results"]
style Q fill:#e1f5fe
style Result fill:#c8e6c9
style Rerank fill:#fff3e0
| Component | Implementation | Features |
|---|---|---|
| BM25 | DuckDB FTS | Exact keyword matching, zero latency |
| Vector | FAISS + sentence-transformers | Semantic understanding, cross-language |
| Hybrid | Weighted fusion | Balance precision and semantics |
| Category | Configuration | Description |
|---|---|---|
| Vector Store | FAISS (IndexFlatIP) |
Inner product index + L2 normalize = cosine similarity |
| Embedding | qwen/qwen3-embedding-4b (2560 dims) |
OpenRouter API, fallback to local 384 dims |
| Dimension Check | β Fail-fast | Error immediately on unexpected API dimensions |
| Chunk (Code) | 50 lines / 10 lines overlap | .py, .js, .go, .rs and 50+ extensions |
| Chunk (Docs) | 256 tokens / 32 tokens overlap | .md, .txt, .rst, .html etc. |
| TopK (hybrid) | kΓ3 (BM25 + Vector each fetch 3x) | Merge β per-file dedup β re-rank |
| Per-file Dedup | β Max 2 chunks per file | Balance recall vs redundancy |
| Max File Size | 1 MB | Larger files auto-skipped |
π Query Flow Diagram (v1.3.3+)
BM25: k*3 = 30 results
Vector: k*3 = 30 results
β Merge + dedup (by source)
~50-60 candidates
β Per-file dedup (max 2 chunks per file)
~35-50 candidates
β Re-rank (LLM subagent)
Return top-10
π Supported File Types (70+)
Code (line-based chunking):
- Python:
.py,.pyw,.pyi,.pyx - JavaScript/TypeScript:
.js,.jsx,.ts,.tsx,.mjs,.cjs - Go:
.go| Rust:.rs| Java:.java| Kotlin:.kt,.kts - C/C++:
.c,.h,.cpp,.cc,.hpp,.hxx - C#:
.cs| Ruby:.rb| PHP:.php| Swift:.swift - Shell:
.sh,.bash,.zsh| SQL:.sql - Config:
.yaml,.yml,.toml,.json,.ini - Web:
.vue,.svelte,.css,.scss - Infra:
.tf,.hcl,.dockerfile,.proto
Docs (token-based chunking):
.md,.markdown,.txt,.rst,.html,.adoc,.org,.tex
Fallback Mechanism:
- Vector dependencies not installed β Auto-fallback to pure BM25
- Vector index doesn't exist β Auto-fallback to pure BM25
Flexible Project Organization
# Three ways to specify a project
./scripts/manage.sh add myproject /path/to/project # By name
./scripts/manage.sh rebuild 45d8fb52 # By ID (8 chars)
./scripts/manage.sh add auto . # Auto-detect
# Claude Code auto-uses current working directory project
# No manual switching requiredMulti-Layer Cache Architecture
- Exact Cache (SQLite): Direct return for exact query matches
- Semantic Cache (FAISS): Cosine similarity matching for similar queries
- LLM Cache: API response caching (reduces repeated calls)
Advantage: Repeated queries return instantly, no recomputation
Long-term Memory + Task Tracking
# Long-term memory (persists across sessions)
memory.set("api_key", "secret_value", project="myproject")
memory.get("api_key")
# Task management
task.add("Implement feature X", priority=10)
task.list(status="in_progress")Professional Management Interface
cd web_ui && ./start.sh # http://localhost:8080- β Real-time log streaming (WebSocket)
- β Interactive search testing
- β Project dashboard
- β Modern dark theme
31 MCP Tools
| Category | Tools |
|---|---|
| RAG | rag.search, answer.generate, answer.accumulated, answer.unified |
| Dual Search | dual.search |
| Project | project.init, project.status |
| Index | index.status, index.rebuild |
| Cache | cache.clear, cache.status |
| Memory | memory.get, memory.set, memory.delete, memory.list, memory.clear |
| Tasks | task.add, task.list, task.update, task.get, task.delete, task.resume, task.current, task.stats |
| Code | code.symbols, code.find_symbol, code.references |
| Search | search.pattern |
| File | file.read, file.list, file.find |
auggie + augment-lite Multi-Engine Orchestration
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β answer.unified (Orchestrator Tool) β
β β
β 1. [minimax-m2.1] Decompose query β sub_queries β
β 2. Return execution plan to Claude β
ββββββββββββββββββββββββ¬βββββββββββββββββββββββββββββββββββββββ
β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β Claude Executes Plan Automatically β
β β
β Step 1: auggie-mcp β semantic_results β
β Step 2: rag.search [minimax re-rank] β rag_results β
β Step 3-N: rag.search (sub-queries) β more_results β
β Step N+1: Merge evidence β [GLM-4.7] β final_answer β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
| Tool | Trigger | Model Calls |
|---|---|---|
answer.accumulated |
Complex questions, previous "don't know" | minimax (decompose) β GLM-4.7 (answer) |
answer.unified |
Need auggie + augment-lite dual engine | minimax (decompose) + auggie + GLM-4.7 |
dual.search |
Only need search results, no answer generation | minimax (re-rank) |
Auto-Rebuild Feature (v1.3.2+)
When dual.search detects >50% of auggie's returned files are missing from augment-lite results, auto-triggers incremental_index rebuild and re-searches:
{
"index_rebuilt": true,
"rebuild_info": {
"files_updated": 15,
"reason": "auggie found files missing from augment-lite index"
}
}| Parameter | Default | Description |
|---|---|---|
auto_rebuild |
true |
Auto-rebuild stale index |
# 1. Clone repository
git clone https://github.com/zoonderkins/augment-lite-mcp.git
cd augment-lite-mcp
# 2. Install dependencies (using uv)
uv venv .venv
source .venv/bin/activate
uv pip install -r requirements.txt
# Or using standard Python
# python3 -m venv .venv && source .venv/bin/activate
# pip install -r requirements.txt
# 3. Configure API Keys
cp .env.example .env
# Edit .env and fill in required API Keys:
# - GLM_API_KEY (from z.ai)
# - MINIMAX_API_KEY (from minimax.io)
# - OPENROUTER_API_KEY (from openrouter.ai/keys) - for Embeddings
# 4. (Optional) Install vector search dependencies (~2GB)
bash scripts/install_vector_deps.sh
# 5. Add project and build index
./scripts/manage.sh add auto .# One-command setup with Claude MCP CLI
claude mcp add --scope user --transport stdio augment-lite \
--env AUGMENT_DB_DIR="$HOME/augment-lite-mcp/data" \
--env GLM_API_KEY="your-glm-api-key" \
--env MINIMAX_API_KEY="your-minimax-api-key" \
--env OPENROUTER_API_KEY="your-openrouter-api-key" \
-- "$HOME/augment-lite-mcp/.venv/bin/python" \
"-u" "$HOME/augment-lite-mcp/mcp_bridge_lazy.py"Edit ~/.claude/config.json:
{
"mcpServers": {
"augment-lite": {
"command": "/absolute/path/to/.venv/bin/python",
"args": ["-u", "/absolute/path/to/mcp_bridge_lazy.py"],
"env": {
"AUGMENT_DB_DIR": "/absolute/path/to/data",
"GLM_API_KEY": "your-glm-api-key",
"MINIMAX_API_KEY": "your-minimax-api-key",
"OPENROUTER_API_KEY": "your-openrouter-api-key"
}
}
}
}Environment Variables:
| Variable | Required | Description |
|---|---|---|
AUGMENT_DB_DIR |
β | Data directory (index, cache, memory) |
GLM_API_KEY |
β | GLM-4.7 API Key (from z.ai) |
MINIMAX_API_KEY |
β | MiniMax-M2.1 API Key (from minimax.io) |
OPENROUTER_API_KEY |
β | OpenRouter API Key (from openrouter.ai/keys) - for Embeddings |
Embedding Note: Uses OpenRouter to call
qwen/qwen3-embedding-4b(2560 dims). IfOPENROUTER_API_KEYnot set, auto-fallback to localall-MiniLM-L6-v2(384 dims).
If using claude-code-proxy local proxy:
{
"env": {
"GLM_LOCAL_BASE_URL": "http://127.0.0.1:8082/v1",
"GLM_LOCAL_API_KEY": "dummy",
"GLM_LOCAL_MODEL_ID": "GLM-4.7",
"MINIMAX_LOCAL_BASE_URL": "http://127.0.0.1:8083/v1",
"MINIMAX_LOCAL_API_KEY": "dummy",
"MINIMAX_LOCAL_MODEL_ID": "MiniMax-M2.1"
}
}Then modify config/models.yaml routes to use glm-local / minimax-local
# In Claude Code
# AI automatically uses augment-lite MCP tools
# Search code
"Help me find the user login handling code"
# Generate answers (with citations)
"How do I configure the database connection?"
# Manage tasks
"Add task: Refactor authentication module"When you first run Claude CLI in a project directory, augment-lite automatically:
1. Project Detection
ββ Auto-identifies current working directory as project
2. Index Building (offline)
ββ BM25 index (DuckDB) - keyword search
ββ Vector index (FAISS) - semantic search (optional)
3. Cache Initialization
ββ Exact cache (SQLite)
ββ Semantic cache (FAISS)
4. Memory Initialization
ββ Long-term memory (SQLite)
Manual Initialization:
./scripts/manage.sh add auto .When you execute any RAG operation (like rag.search), augment-lite handles everything automatically:
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β User executes rag.search β
βββββββββββββββββββββββββββ¬ββββββββββββββββββββββββββββββββ
β
βΌ
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β resolve_auto_project() β
β Priority: 1.Dir name match 2.Path match 3.Active 4.None β
βββββββββββββββββββββββββββ¬ββββββββββββββββββββββββββββββββ
β
βββββββββββββ΄ββββββββββββ
β β
Project registered Project not registered
β β
β βΌ
β βββββββββββββββββββββββββββ
β β AUTO-INIT β
β β - Sanitize dir name β
β β - Register to projects β
β β - Set as active β
β βββββββββββββ¬ββββββββββββββ
β β
ββββββββββββ¬βββββββββββ
β
βΌ
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β auto_index_if_needed() β
β Detect file changes β
βββββββββββββββββββββββββββ¬ββββββββββββββββββββββββββββββββ
β
βββββββββββββ΄ββββββββββββ
β β
Has index state No index state
β β
βΌ βΌ
ββββββββββββββββββ ββββββββββββββββββ
β Incremental β β Full build β
β Only process β β All files as β
β changed files β β "added" β
β (mtime+MD5) β β β
ββββββββββ¬ββββββββ ββββββββββ¬ββββββββ
β β
βββββββββββββ¬ββββββββββββ
β
βΌ
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β Execute Search β
β BM25 + Vector hybrid β LLM filtering β Return results β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
| Feature | Description |
|---|---|
| Zero Config | No manual project.init needed, first search auto-initializes |
| Zero Maintenance | Auto-detect file changes, incremental index updates |
| Smart Sanitization | Special chars like . in dir names auto-replaced with - |
| Auto Activation | New projects auto-set as active, no manual switching |
# Traditional way (still supported)
project.init β index.rebuild β rag.search
# New way (recommended)
rag.search # Auto-handles init + index| Tool | Description | Usage |
|---|---|---|
rag.search |
BM25 + vector hybrid search | Search code snippets |
answer.generate |
Generate answers from retrieval | Answers with citations |
answer.accumulated |
Multi-round evidence accumulation | Complex questions, avoid "don't know" |
answer.unified |
auggie + augment-lite orchestration | Returns execution plan for Claude |
dual.search |
Dual-engine search | Local RAG + auggie hint |
index.rebuild |
Rebuild project index | Use when index corrupted |
index.status |
Check index status | View index health |
| Tool | Description | Example |
|---|---|---|
code.symbols |
Get code symbol overview | List classes, functions, methods |
code.find_symbol |
Find symbol definition | Find MyClass definition location |
code.references |
Find symbol references (AST) | Find all usages of my_func |
search.pattern |
Regex pattern search | def.*search matching |
file.read |
Read file content | Supports line ranges |
file.list |
List directory contents | Supports glob filters |
file.find |
Find files | **/*.py patterns |
Supported Languages (v1.3.0+, Tree-sitter):
Python, JavaScript, TypeScript, Go, Rust, Bash,
JSON, YAML, HTML, CSS, HCL (Terraform), TOML
| Tool | Description |
|---|---|
memory.get/set/delete/list |
Long-term memory management |
task.add/list/update/delete |
Task tracking |
project.init/status |
Project management |
cache.clear/status |
Cache management |
AI proactively auto-stores Memory in these situations:
| Trigger Condition | Suggested Key | Example Content |
|---|---|---|
| First interaction | project_overview |
Project name, version, purpose, core modules |
| Code style detected | code_style |
ruff, 100 chars, Python 3.12+ |
| Read README/docs | suggested_commands |
setup, build, test commands |
| User explains architecture | architecture_decisions |
Microservices, GraphQL, Redis cache |
| User mentions deployment | deployment_notes |
Docker, K8s, environment variables |
| Bug pattern found | known_issues |
Common errors, workarounds |
Standard Memory Keys:
project_overview # Project overview
code_style # Code style
suggested_commands # Common commands
architecture_decisions # Architecture decisions
api_conventions # API conventions
deployment_notes # Deployment notes
known_issues # Known issues
task_completion_checklist # Completion checklist
Save Tokens + Improve Semantic Accuracy
augment-lite can work collaboratively with Auggie MCP:
ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β Claude Code β
βββββββββββββββββββ¬βββββββββββββββββββββββββββββββββββββββββ
β
ββββββββββββββ΄βββββββββββββ
β β
ββββββΌβββββββββ ββββββΌβββββββββ
β augment-liteβ β auggie-mcp β
β (BM25 RAG) β β (Semantic) β
βββββββββββββββ βββββββββββββββ
β β
ββββββββββββ¬βββββββββββββββ
β
βββββββββΌββββββββ
β Merge Resultsβ
β Dedup + Sort β
βββββββββββββββββ
# 1. Install auggie CLI
npm install -g @augmentcode/auggie@latest
# 2. Login (requires Augment Code account)
auggie login
# 3. Add to Claude Code
claude mcp add-json auggie-mcp --scope user '{"type":"stdio","command":"auggie","args":["--mcp"]}'| Mode | Description | Token Consumption |
|---|---|---|
| augment-lite only | BM25 + Vector local search | Low |
| auggie-mcp only | Augment semantic engine | Lowest |
| Hybrid mode (Recommended) | augment-lite initial filter + auggie refinement | Medium |
1. User query "find auth module"
2. augment-lite rag.search β Get 20 candidates
3. auggie-mcp codebase-retrieval β Semantic sort
4. Claude merges results β Return most relevant 5
Fallback without Auggie: augment-lite uses built-in LLM filtering (GLM-4.7/MiniMax), can work independently
| Module | Function | Description |
|---|---|---|
prompt_injection |
Prompt injection detection | Prevent instruction override, role hijacking, jailbreak |
pii_detection |
PII/sensitive data detection | Email, phone, SSN, API Key, JWT |
code_security |
Code security scanning | SQL injection, XSS, command injection, OWASP Top 10 |
hallucination |
Hallucination detection | Verify answers are based on provided evidence |
context_grounding |
Context grounding | Ensure answers don't exceed provided context |
schema_validation |
Output validation | JSON schema validation, MCP output format |
| Variable | Description | Source |
|---|---|---|
GLM_API_KEY |
GLM-4.7 API Key | z.ai |
MINIMAX_API_KEY |
MiniMax-M2.1 API Key | minimax.io |
| Variable | Description | Default |
|---|---|---|
AUGMENT_DB_DIR |
Data directory | ./data |
If using claude-code-proxy:
| Variable | Description |
|---|---|
GLM_LOCAL_BASE_URL |
http://127.0.0.1:8082/v1 |
GLM_LOCAL_API_KEY |
dummy |
MINIMAX_LOCAL_BASE_URL |
http://127.0.0.1:8083/v1 |
MINIMAX_LOCAL_API_KEY |
dummy |
ββββββββββββββββββββββββββββββββββββββββββββββββ
β Claude Code (AI Assistant) β
βββββββββββββββββββ¬βββββββββββββββββββββββββββββ
β MCP Protocol
βββββββββββββββββββΌβββββββββββββββββββββββββββββ
β mcp_bridge_lazy.py (31 Tools) β
βββββββββββββββββββ¬βββββββββββββββββββββββββββββ
β
ββββββββββββββΌβββββββββββββ
β β β
ββββββΌββββββ ββββΌβββββββ βββΌβββββββββ
β Retrievalβ β Cache β β Memory β
β (BM25+ β β (3-Layerβ β (SQLite) β
β Vector) β β Cache) β ββββββββββββ
ββββββ¬ββββββ βββββββββββ
β
ββββββΌββββββββββββββββββββββββββββββββββ
β Layer 1: API/Local Embeddings β
β - OpenRouter: qwen3-embedding-4b β
β - Fallback: sentence-transformers β
β - BM25 + FAISS hybrid search β
β - 50 candidates β
ββββββ¬ββββββββββββββββββββββββββββββββββ
β
ββββββΌββββββββββββββββββββββββββββββββββ
β Layer 2: Remote LLM Re-ranking β
β - GLM-4.7 / MiniMax-M2.1 (native) β
β - Smart filtering + deduplication β
β - Final 8 results β
ββββββββββββββββββββββββββββββββββββββββ
| Provider | Endpoint | Context | Max Output |
|---|---|---|---|
| glm-4.7 | api.z.ai/api/anthropic |
200K | 128K |
| minimax-m2.1 | api.minimax.io/anthropic |
200K | - |
| Route | Provider | Max Output | Trigger |
|---|---|---|---|
small-fast |
minimax-m2.1 | 2048 | lookup, small_fix |
general |
glm-4.7 | 4096 | general tasks |
reason-large |
glm-4.7 | 8192 | refactor, reason |
big-mid |
glm-4.7 | 8192 | tokens > 200K |
long-context |
glm-4.7 | 8192 | tokens > 400K |
ultra-long-context |
glm-4.7 | 16384 | ultra-long context |
fast-reasoning |
minimax-m2.1 | 4096 | fast reasoning |
- Auto-incremental indexing (zero maintenance)
- Dual-layer retrieval (local + remote)
- Multi-project management (name/ID/auto)
- Three-layer caching (exact + semantic + provider)
- Long-term memory (global/project scope)
- Task management (structured tracking)
- Web UI (FastAPI + WebSocket)
- MCP protocol compliance (28 tools)
- AI auto-discovery (server instructions)
- Gitignore filtering
- Model-specific system prompts
- Dynamic token limits
- Guardrails (evidence citation)
- Modern Guardrails (v1.2.0)
- Prompt Injection Detection
- PII/API Key Detection
- Code Security Scanning (OWASP)
- Hallucination Detection
- Context Grounding Validation
- Output Schema Validation
| Approach | Functionality | Status |
|---|---|---|
| Tree-sitter | AST structure parsing (12 languages) | β v1.3.0 implemented |
| Modify Symbol Tools | rename, edit_symbol_body | π§ v1.4.0 planned |
| LSP Bridge | Complete semantic analysis | π§ v1.5.0 planned |
| Auggie MCP | External semantic engine | π§ Optional integration |
Completed (v1.3.0):
- Tree-sitter Integration: 12 language AST parsing
- Python, JavaScript, TypeScript, Go, Rust, Bash
- JSON, YAML, HTML, CSS, HCL (Terraform), TOML
- AST-based References:
code.referencesuses Tree-sitter for precise location
v1.4.0 Plan (P85):
- Modify Symbol Tools (Serena core gap)
code.rename_symbol- Scope-aware safe renamingcode.edit_symbol_body- Precise symbol content replacementcode.insert_before_symbol/code.insert_after_symbol- Initial approach: AST-based Python-only version
v1.5.0 Plan (P80):
- LSP β MCP Bridge: Complete semantic analysis
- rename (cross-file safe refactoring)
- diagnostics (lint/errors)
- code actions
- Approach: Integrate mcp-language-server or build LSP client wrapper
Research (P70):
- Auggie MCP Deep Integration: Evaluate as LSP alternative
- Install:
npm install -g @augmentcode/auggie@latest && auggie login - Config:
claude mcp add-json auggie-mcp --scope user '{"type":"stdio","command":"auggie","args":["--mcp"]}'
- Install:
- Multi-language embeddings (multilingual-e5-large)
- Code-specific embeddings (CodeBERT, UniXcoder)
- GraphRAG integration (code dependency graphs)
- Incremental vector index updates
- Cloud deployment options (Docker Compose)
- VSCode extension (alternative to MCP)
- Monitoring dashboard (metrics, usage stats)
- Plugin system (custom tools)
| Characteristic | Description |
|---|---|
| Indexing | DuckDB BM25 batch indexing, supports large codebases |
| Incremental | Only update changed files, avoid full rebuild |
| Local First | BM25+Vector computed locally, no API latency |
| LLM Filtering | Only candidates sent to LLM, reduced token consumption |
| Cache | Repeated queries return directly, no computation overhead |
Want to understand how augment-lite-mcp differs from other solutions?
- vs Anthropic @modelcontextprotocol/context: View comparison
- vs acemcp: View comparison
- vs Augment Code: View comparison
- vs Qdrant/Weaviate: View comparison
- Vector Models Selection Guide: View details
-
acemcp by @wxxedu
- π‘ Auto-incremental indexing implementation
- π‘ Zero-maintenance philosophy
- π‘ Web UI design inspiration
-
Augment Code (Proprietary)
- π‘ Context Engine architecture insights
- π‘ Two-stage retrieval (local + remote) concept
-
@modelcontextprotocol/context by Anthropic
- π‘ MCP protocol standard reference
- π‘ Simple and efficient file access design
-
sentence-transformers by Hugging Face
- all-MiniLM-L6-v2 embedding model
- Local, free, high-quality
-
- Multi-model aggregation platform
- 300+ models unified API
-
DuckDB - Embedded SQL database
-
FAISS (Meta) - Vector similarity search
-
FastAPI - Modern web framework
-
Claude Code - MCP protocol and development tools
# docs/ directory contains full documentation (added to .gitignore)
# Read locally or generate online if neededdocs/guides/- Usage guides- MCP Setup, Multi-Project, Vector Search, Cache, Memory, Tasks
docs/features/- Feature descriptionsdocs/core/- Architecture and technical overviewdocs/bugfixes/- Bug fix records
init/specs/- Technical specificationsinit/guidelines/- Coding standards, naming conventions, documentation guideinit/workflows/- Release, bugfix, feature development workflows
# Quick environment check
./health_check.sh
# Full test suite
python tests/run_all_tests.py
# Individual tests
python tests/test_high_priority_apis.pySee TESTING.md for details
Contributions welcome! Please follow this process:
- Fork this repository
- Create a feature branch (
git checkout -b feature/AmazingFeature) - Commit changes (
git commit -m 'Add some AmazingFeature') - Push to branch (
git push origin feature/AmazingFeature) - Open a Pull Request
See init/workflows/RELEASE_WORKFLOW.md and init/guidelines/CODING_STANDARDS.md
This project is licensed under the MIT License - see the LICENSE file for details
- Repository: https://github.com/zoonderkins/augment-lite-mcp
- Issues: https://github.com/zoonderkins/augment-lite-mcp/issues
- Changelog: CHANGELOG.md
- MCP Protocol: https://github.com/anthropics/mcp
- GitHub Issues: Report bugs or feature requests
- Discussions: Ask questions or share experiences
Made with β€οΈ by the community
Inspired by acemcp, Augment Code, and the open-source AI community