Director of Platform Engineering. Building LLMKube, open source K8s operator for local LLM inference.
- Seattle, WA
- https://defilan.com
- @defilan
Highlights
- Pro
Pinned Loading
-
defilantech/LLMKube
defilantech/LLMKube PublicKubernetes operator for GPU-accelerated LLM inference - air-gapped, edge-native, production-ready
-
defilantech/infercost
defilantech/infercost PublicKubernetes-native cost intelligence for on-premises AI inference. Computes true cost-per-token from GPU amortization, electricity, and real power draw.
Go 2
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.





