Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
20 changes: 20 additions & 0 deletions packages/app/public/logos/weka.svg
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
1 change: 1 addition & 0 deletions packages/app/src/components/page-content.tsx
Original file line number Diff line number Diff line change
Expand Up @@ -311,6 +311,7 @@ export function PageContent({ initialTab = 'inference' }: { initialTab?: string
'Crusoe',
'TensorWave',
'SGLang',
'WEKA',
].includes(q.org),
)}
overrides={{
Expand Down
7 changes: 7 additions & 0 deletions packages/app/src/components/quotes/quotes-data.ts
Original file line number Diff line number Diff line change
Expand Up @@ -237,4 +237,11 @@ export const QUOTES: Quote[] = [
org: 'Verda',
logo: 'verda.svg',
},
{
text: "InferenceX\u2122 ensembles precisely that \u2014 open, reproducible benchmarks that are continuously updated as xPU accelerators (GPUs/TPUs/LPUs), memory, storage, and software stacks evolve. I'm excited to see the InferenceX benchmarking roadmap include agentic coding workloads that stress CPU KV Cache offloading & soon NVMe KV Cache offloading from xPUs. As WEKA helps scale the Memory Wall by building the KV Cache infrastructure that feeds these xPUs, having this level of visibility into inference performance helps the entire ecosystem make smarter decisions about where to invest.",
name: 'Val Bercovici',
title: 'Chief AI Officer, WEKA',
org: 'WEKA',
logo: 'weka.svg',
},
];