Nexus 1.7 Core is a spine-first, model-aware AI orchestration system. Instead of sending every prompt to one large model, Nexus breaks the problem down, decides what kind of task it is, and routes it to the model that is most likely to do the best job.
This project focuses on control, cost efficiency, and reliability, not just raw model power.
Most AI systems:
- Use one model for everything
- Overpay for simple tasks
- Struggle with consistency on complex requests
- Have no internal quality checks
Nexus flips that approach.
It treats models as specialized tools, not general-purpose magic.
-
Spine-first architecture One primary model (“spine”) is responsible for generating the final output.
-
Helpers only when needed Secondary models (“micro-minds”) are used strictly for validation, not generation.
-
Confidence-driven execution If the system is confident, it stays fast and cheap. If not, it slows down and double-checks.
-
Explicit, debuggable flow Every step is logged and traceable.
When a prompt is submitted, Nexus runs through the following pipeline:
A fast, low-cost model analyzes the prompt and determines:
-
Task type:
- coding
- writing
- reasoning
- chat
-
Complexity:
- trivial
- simple
- moderate
- complex
This step exists solely to route the task correctly.
Based on classification, Nexus selects a primary spine model.
Each task type has a purpose-built model:
- Coding → code-optimized models
- Writing → prose and documentation models
- Reasoning → deep analytical models
- Chat → fast conversational models
Token limits and temperature are adjusted automatically based on complexity.
Before generating the final response, Nexus evaluates:
- How confident the system is that the spine can handle this task alone
- Which validators (if any) are actually necessary
This produces:
- A confidence score between 0 and 1
- A short list of optional micro-minds
If confidence is high enough, validation is skipped entirely.
If enabled, Nexus performs a dedicated reasoning pass before generation.
This phase:
- Breaks down the task step-by-step
- Identifies risks and edge cases early
- Improves output quality on complex tasks
Modes:
- Disabled – fastest, lowest cost
- Standard – balanced reasoning
- Extended – deep analysis for difficult problems
The output of this phase is used as internal context, not shown to the user.
The selected spine model generates the actual response.
Key rules:
- One spine produces the final answer
- System prompts enforce formatting and behavior
- Code output follows strict file boundaries
- Streaming is supported for real-time UX
This keeps responsibility clear and avoids model conflicts.
If confidence is below the configured threshold, Nexus invokes helper models.
Each micro-mind has a single responsibility:
-
Logic Validator Checks for contradictions or flawed reasoning
-
Edge Case Detector Looks for missing cases or unhandled scenarios
-
Sanity Checker Ensures the output is reasonable and coherent
-
Syntax Scanner (code only) Flags syntax and structural issues
Micro-minds do not rewrite the output.
They only return short notes.
If any validator reports a critical issue:
- The spine is called again
- The issues are explicitly injected into the prompt
- A corrected response is generated
Non-critical feedback is logged but does not trigger regeneration.
✔ Faster responses for simple prompts ✔ Lower costs on average workloads ✔ Higher reliability on complex tasks ✔ Clear ownership of output ✔ No “model soup” or hidden behavior
This system scales well as new models are added because routing logic stays the same.
Nexus includes a terminal-based testing console with:
- Live conversations
- Full execution logs
- Model and confidence visibility
- Token and timing statistics
Built-in commands:
helpconfigclearstatsexit
This makes experimentation and debugging straightforward.
All behavior is controlled through a single config object:
- Thinking depth
- Output token limits
- Confidence thresholds
- Maximum helper count
- Verbose logging
- Micro-mind enable/disable
Nothing is hidden or implicit.
- Node.js 18+
- OpenRouter API key
npm install openai- TypeScript runtime (
tsx)
This project is intended for:
- Developers building AI platforms
- People experimenting with multi-model systems
- Anyone who wants control over AI behavior
- Systems where cost and correctness both matter