diff --git a/packages/app/public/logos/weka.svg b/packages/app/public/logos/weka.svg new file mode 100644 index 0000000..c24d05e --- /dev/null +++ b/packages/app/public/logos/weka.svg @@ -0,0 +1,20 @@ + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/packages/app/src/components/page-content.tsx b/packages/app/src/components/page-content.tsx index 4d57b03..5216b43 100644 --- a/packages/app/src/components/page-content.tsx +++ b/packages/app/src/components/page-content.tsx @@ -311,6 +311,7 @@ export function PageContent({ initialTab = 'inference' }: { initialTab?: string 'Crusoe', 'TensorWave', 'SGLang', + 'WEKA', ].includes(q.org), )} overrides={{ diff --git a/packages/app/src/components/quotes/quotes-data.ts b/packages/app/src/components/quotes/quotes-data.ts index c49fe70..b9187e7 100644 --- a/packages/app/src/components/quotes/quotes-data.ts +++ b/packages/app/src/components/quotes/quotes-data.ts @@ -237,4 +237,11 @@ export const QUOTES: Quote[] = [ org: 'Verda', logo: 'verda.svg', }, + { + text: "InferenceX\u2122 ensembles precisely that \u2014 open, reproducible benchmarks that are continuously updated as xPU accelerators (GPUs/TPUs/LPUs), memory, storage, and software stacks evolve. I'm excited to see the InferenceX benchmarking roadmap include agentic coding workloads that stress CPU KV Cache offloading & soon NVMe KV Cache offloading from xPUs. As WEKA helps scale the Memory Wall by building the KV Cache infrastructure that feeds these xPUs, having this level of visibility into inference performance helps the entire ecosystem make smarter decisions about where to invest.", + name: 'Val Bercovici', + title: 'Chief AI Officer, WEKA', + org: 'WEKA', + logo: 'weka.svg', + }, ];