Skip to content

Commit 2348268

Browse files
committed
fix(ci): pin onnxruntime>=1.20.0 to ensure INT4 quantization support
- Specify minimum version 1.20.0 which has stable matmul_4bits_quantizer - Install optimum without [onnxruntime] extra to avoid version downgrades - Add package version diagnostics for debugging - Fail fast if INT4 quantization is not available (instead of silent fallback)
1 parent 757d3ed commit 2348268

File tree

1 file changed

+6
-3
lines changed

1 file changed

+6
-3
lines changed

.github/workflows/build-wasm.yml

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -163,10 +163,13 @@ jobs:
163163
run: |
164164
echo "::group::Installing Python ML dependencies"
165165
pip install --no-cache-dir torch transformers
166-
pip install --no-cache-dir --upgrade onnx onnxruntime onnxruntime-tools
167-
pip install --no-cache-dir optimum[onnxruntime]
166+
pip install --no-cache-dir "onnx>=1.15.0" "onnxruntime>=1.20.0"
167+
pip install --no-cache-dir optimum
168+
echo "Installed packages:"
169+
pip list | grep -E "(onnx|optimum|torch)"
170+
echo ""
168171
python3 -c "import onnxruntime; print(f'ONNX Runtime version: {onnxruntime.__version__}')"
169-
python3 -c "from onnxruntime.quantization import matmul_4bits_quantizer; print('INT4 quantization available')" || echo "INT4 quantization not available - will use FP32 fallback"
172+
python3 -c "from onnxruntime.quantization import matmul_4bits_quantizer; print('INT4 quantization available')"
170173
echo "::endgroup::"
171174
172175
- name: Install dependencies

0 commit comments

Comments
 (0)