CPU backends precisions support #18286
Unanswered
truecoder34
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi all,
Does CPU backends supported in ggml (ZDNN, ZenDNN, BLAS, KleidiAI) only able to run inference of models with F16/B16 and F32 precision only, and any quantized models will be run with Naive CPU backend?
Is this limitation actual for all CPU Backends excluding Naive CPU one ?
Thank you in advance .
Beta Was this translation helpful? Give feedback.
All reactions