You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Complete reference for common mini-a parameters. Parameters are set as param=value arguments, by using the corresponding OAF_*/MINI_A_* environment variables where supported, or through saved model definitions.
mini-a param=value
or
export MINI_A_PARAM=value
1. Model Configuration
Parameter
Default
Description
model
-
LLM model configuration in SLON/JSON style (e.g., (type: openai, model: gpt-5-mini, key: '...'))
modellc
-
Lighter model for simple tasks (dual-model); set via OAF_LC_MODEL env var
modelval
-
Runtime override for the validation model configuration; same format as OAF_VAL_MODEL
modellock
auto
Force model-tier selection: main, lc, or auto
lccontextlimit
0
Escalate from low-cost model to main model when context tokens reach this threshold (0 disables)
deescalate
3
Consecutive successful steps required before switching back to the low-cost model after escalation
lcescalatedefer
true
Defer LC-to-main escalation by one step when LC confidence remains high
lcbudget
0
Maximum total LC token budget before permanently switching to the main model (0 = unlimited)
llmcomplexity
false
Use a quick LC validation call for medium-complexity routing heuristics
modelstrategy
default
Model orchestration profile: default (LC-first with escalation) or advisor (LC executes, main model consulted selectively for difficult steps)
advisormaxuses
2
Maximum advisor consultations per run when modelstrategy=advisor
advisorcooldownsteps
2
Minimum step distance between advisor consultations when modelstrategy=advisor
maxtokens
-
Maximum output tokens
rpm
-
Requests per minute limit
tpm
-
Tokens per minute limit
2. Core Parameters
Parameter
Default
Description
goal
-
The task/goal for the agent to accomplish
verbose
false
Print more detailed runtime logs
useshell
false
Enable shell command execution
chatbotmode
false
Pure chat mode without tools
maxsteps
15
Maximum number of agent steps
rtm
-
Legacy alias for rpm
format
-
Output format (e.g., json, yaml, markdown)
youare
-
Custom system persona/identity
chatyouare
-
Chatbot-mode persona override when chatbotmode=true
rules
-
Additional rules for the agent
knowledge
-
Knowledge base content or file path
conversation
-
Conversation history file to preload at startup
state
-
Structured initial state data (JSON/SLON) provided to the agent
raw
false
Return raw output without extra formatting
showthinking
false
Surface XML-tagged model thinking blocks as thought logs
debug
false
Enable debug logging
debugfile
-
Write debug output as NDJSON to a file (implies debug=true)
debugch
-
SLON channel definition for main-model LLM traffic — see [Channels]({{ '/channels'
debuglcch
-
SLON channel definition for low-cost-model LLM traffic — see [Channels]({{ '/channels'
debugvalch
-
SLON channel definition for validation-model traffic when llmcomplexity=true — see [Channels]({{ '/channels'
outfile
-
Save final answer to file
outfileall
-
Deep research only: save full cycle history/verdicts/learnings to file
outputfile
-
Alternate key for outfile, mainly used by plan conversion flows
extracommands
-
Comma-separated extra directories for custom slash command templates
extraskills
-
Comma-separated extra directories for custom skills
extrahooks
-
Comma-separated extra directories for hook definitions
secpass
-
Password used to open OpenAF sBucket model secrets
auditch
-
SLON channel definition for agent interaction audit logs — see [Channels]({{ '/channels'
toollog
-
SLON channel definition for dedicated MCP tool input/output logs — see [Channels]({{ '/channels'
metricsch
-
SLON/JSON channel definition for recording periodic Mini-A metrics snapshots (e.g. (name: 'mini-a-metrics', type: 'mvs', options: (file: '/tmp/mini-a-metrics.db'))). Supports optional period, some, and noDate fields — see [Channels]({{ '/channels'
3. MCP Configuration
Parameter
Default
Description
mcp
-
MCP servers to load (comma-separated)
mcpproxy
false
Enable MCP proxy mode
mcpproxythreshold
0
Byte threshold to spill large proxy results to temp files (0 disables spilling)
mcpproxytoon
false
Serialize spilled proxy object/array payloads as TOON text when proxy spilling is enabled
mcpprogcall
false
Start localhost programmatic MCP tool-call bridge for scripts
mcpprogcallport
0
Programmatic MCP bridge port (0 auto-selects)
mcpprogcallmaxbytes
4096
Max inline JSON response size before returning a stored resultId
mcpprogcallresultttl
600
TTL (seconds) for oversized stored results served by the MCP bridge
mcpprogcalltools
""
Optional comma-separated tool allowlist exposed by the MCP bridge
mcpprogcallbatchmax
10
Maximum calls accepted by one bridge batch request
mcpdynamic
false
Allow dynamic MCP discovery
mcplazy
false
Lazy-load MCP servers
mcpurl
-
Remote MCP server URL
nosetmcpwd
false
Do not set the default MCP command working directory to the mini-a install path
4. Tool Configuration
Parameter
Default
Description
useutils
false
Enable Mini Utils Tool utilities (init, filesystemQuery, filesystemModify, markdownFiles)
useskills
false
Expose skill operations in Mini Utils Tool (requires useutils=true)
utilsroot
-
Root path used by Mini Utils file operations
mini-a-docs
false
If true and utilsroot is not set, automatically uses the mini-a oPack docs path as utilsroot
miniadocs
false
Alias for mini-a-docs
usetools
false
Enable tool usage
usejsontool
false
Register a compatibility json tool for models that sometimes emit json tool calls
libs
-
Additional library paths to load
toolcachettl
600000
Default cache TTL in milliseconds for MCP tool results
utilsallow
-
Comma-separated allowlist of Mini Utils Tool names to expose when useutils=true
utilsdeny
-
Comma-separated denylist of Mini Utils Tool names to hide when useutils=true (applied after utilsallow)
5. Context Management
Parameter
Default
Description
maxcontext
-
Maximum context window tokens
maxcontent
-
Alias for maxcontext
maxtokens
-
Maximum response tokens
autocompact
true
Auto-compact when context is full
6. Deep Research
Parameter
Default
Description
deepresearch
false
Enable iterative research-validate-learn cycles
validationgoal
-
Validation criteria for deep research output quality (inline text or file path)
valgoal
-
Alias for validationgoal
vmodel
-
Optional dedicated validation model used in deep-research scoring
modelval
-
Per-run validation-model override using the same SLON/JSON definition accepted by OAF_VAL_MODEL
maxcycles
3
Maximum number of deep-research cycles
validationthreshold
PASS
Validation verdict or score rule required to stop iterating
persistlearnings
true
Carry learnings from failed validation cycles into the next cycle
7. Planning
Parameter
Default
Description
useplanning
false
Enable agent planning
forceplanning
false
Force planning even when heuristics would normally skip it
Path to an agent markdown file, or inline markdown with YAML frontmatter
agentfile
-
Backward-compatible alias for agent
See the [Agent Files]({{ '/agents' | relative_url }}) page for the complete reference: all frontmatter keys, tool entry types, mini-a: overrides, relative file resolution, precedence rules, and a full annotated example.
10b. Conversation History
Parameter
Default
Description
historykeep
false
Save console conversations to ~/.openaf-mini-a/history for future resumption
historykeepperiod
-
Delete kept conversation files older than this many minutes
historykeepcount
-
Keep only the newest N kept conversation files
10b. Working Memory
Enable a structured, scoped working memory subsystem that the agent maintains automatically across tool calls and runs.
Parameter
Default
Description
usememory
false
Enable the working memory subsystem
memoryuser
false
Convenience preset that enables file-backed session + global memory under ~/.openaf-mini-a/, auto-promotes facts,decisions,summaries, and enables a 30-day stale sweep
memoryscope
both
Scope of memory: session (current run only), global (shared across sessions), or both
memorych
-
SLON/JSON channel definition for global memory persistence (e.g. file, Redis)
memorysessionch
-
SLON/JSON channel definition for session memory (falls back to memorych if omitted)
memorysessionid
agent ID
Key namespace used for session memory in the channel
memorymaxpersection
80
Maximum entries per memory section before compaction
memorymaxentries
500
Hard cap on total entries across all sections
memorycompactevery
8
Number of appends between automatic compaction passes
Comma-separated sections to auto-promote from session memory to global memory at session end
memorystaledays
0
Days without re-confirmation before a global entry is marked stale; memoryuser=true sets this to 30
memoryinject
summary
summary injects only per-section counts and enables on-demand memory_search; full embeds the full compact memory snapshot in every step
Memory is organized into 8 sections: facts, evidence, decisions, risks, openQuestions, hypotheses, artifacts, summaries. The agent appends entries automatically at significant events (tool calls, plan critiques, validation results, final answers).
When both memorych and memorysessionch are configured, writes under memoryscope=both default to the session store first. Auto-promotion then refreshes or appends selected knowledge into the global store at session end.
10c. Adaptive Routing
A rule-based routing layer that selects how each tool action is dispatched. When disabled, mini-a uses legacy behavior.
Parameter
Default
Description
adaptiverouting
false
Enable rule-based adaptive tool routing
routerorder
-
Preferred route execution order (comma-separated route types)
routerallow
-
Allowlist of route types the router may select (comma-separated)
routerdeny
-
Denylist of route types the router must never select (comma-separated)
routerproxythreshold
-
Byte threshold to prefer the proxy route for large payloads (falls back to mcpproxythreshold when omitted)
Route selection is based on intent hints (read/write, payload size, latency sensitivity, risk level, structured output preference) and historical route success. Fallback chains are attempted in order; each failed attempt is appended to an errorTrail. Route decisions are visible in debug/audit output as [ROUTE ...] records when debug=true.
11. Web Interface
Parameter
Default
Description
onport
-
Port for web UI (enables web mode)
maxpromptchars
120000
Maximum accepted prompt size for incoming web /prompt requests
12. Knowledge & Persona
Parameter
Default
Description
knowledge
-
Knowledge base content or file
youare
-
Agent persona/identity description
chatyouare
-
Chatbot-mode persona override
rules
-
Behavioral rules for the agent
13. Rate Limiting
Parameter
Default
Description
rpm
-
Requests per minute limit
tpm
-
Tokens per minute limit
rtm
-
Legacy alias for rpm
14. Docker Environment Variables
Variable
Maps to
OAF_MODEL
model
OAF_LC_MODEL
modellc
OAF_VAL_MODEL
modelval — dedicated validation model for deep research scoring
OAF_MINI_A_NOJSONPROMPT
Force text prompt mode for main model; Gemini main models auto-enable this behavior when unset
OAF_MINI_A_LCNOJSONPROMPT
Force text prompt mode for low-cost model (set explicitly for Gemini low-cost models)
OAF_MINI_A_CON_HIST_SIZE
Maximum console history size (defaults to JLine's default)
OAF_MINI_A_LIBS
Comma-separated library paths to load automatically at startup
MINI_A_GOAL
goal
MINI_A_PORT
onport
OAF_MODEL / OAF_LC_MODELkey field
Provider API credential (recommended)
GITHUB_TOKEN
GitHub Models token (optional when key is provided in model config)
15. Console Commands Reference
Command
Description
/help
Show available commands
/model
Show current model info
/models
Show all configured model tiers (main, LC, validation) with provider and source
/compact [n]
Compact older history while keeping up to latest n exchanges (default 6)
/summarize [n]
Summarize older history while keeping up to latest n exchanges (default 6)
`/context [llm
analyze]`
/reset
Reset conversation
/last [md]
Reprint the previous final answer (raw markdown with md)
/save <path>
Save the previous final answer to a file
/stats [mode] [out=file.json]
Show session metrics (summary/detailed/tools) and optionally export JSON
/history [n]
Show the latest user goals from conversation history
/exit
Exit mini-a
/clear
Reset conversation history and accumulated metrics
Internet-focused MCP/tool preset with docs-aware utils and proxy aggregation
news
News-focused MCP preset (web + rss + time, proxy enabled)
poweruser
Shell + utils + tools with proxy tuning, history retention, LC validation, and docs-aware defaults
web
Browser UI preset with MCP tools enabled
webfull
Full web UI preset with history/attachments, proxy tuning, charts/diagrams/maps, and richer rendering options
User custom presets can be defined in ~/.openaf-mini-a_modes.yaml. They are merged with built-ins from mini-a-modes.yaml, and user definitions take precedence.
Channels
mini-a uses OpenAF channels as the pluggable storage backend for audit logs, tool logs, and debug traffic. The channel-accepting parameters (auditch, toollog, debugch, debuglcch, debugvalch, metricsch) each accept a SLON definition that specifies the backend, channel name, and options.
See the [Channels]({{ '/channels' | relative_url }}) reference for the complete definition format, every supported backend (file, MVS, Redis, S3, and more), and ready-to-use query examples.
Advanced Topics
For power-user configuration, deployment patterns, and provider-specific guides, see the [Advanced]({{ '/advanced' | relative_url }}) page. Topics covered: