On-device shell command generator for macOS Tahoe. Uses Apple's 3B model with dynamic few-shot retrieval from 21k tldr examples.
-
Updated
Apr 19, 2026 - Python
On-device shell command generator for macOS Tahoe. Uses Apple's 3B model with dynamic few-shot retrieval from 21k tldr examples.
Agentic Android Open Source Project (AAOSP) — Android fork with native LLM system service, MCP-aware apps, and an agent-driven launcher. On-device Qwen 2.5 via llama.cpp. Apps declare tools in their manifest. The OS runs the model.
KnoLo Core is a local-first knowledge base engine built for small language models (LLMs). It packages your documents into a compact .knolo file and enables fully deterministic querying — no embeddings, no vector databases, no cloud services required. Designed for on-device and edge LLM deployments.
High-performance Android SDK for on-device LLM inference (GGUF). Privacy-focused, offline-first, and powered by llama.cpp with a clean Kotlin Coroutines API.
Documentation for MobileTransformers - a lightweight, modular framework based on ONNX Runtime for running and adapting large language models (LLMs) directly on mobile and edge devices. It supports on-device fine-tuning (PEFT), efficient inference, quantization, weight merging, and direct inference from merged models.
llama.cpp packaged for AOSP — Android.bp build rules, JNI bridge, and Qwen 2.5 model download scripts for on-device inference
📱 手机端 AI 操作系统全景知识库 — 334+ 篇深度页面,覆盖端侧大模型、AI Agent、芯片适配、推理优化 | 自动更新
Add a description, image, and links to the on-device-llm topic page so that developers can more easily learn about it.
To associate your repository with the on-device-llm topic, visit your repo's landing page and select "manage topics."