From 0ee8fb9560b9412ab635a213cf17cff9b883a8aa Mon Sep 17 00:00:00 2001 From: Arav Agarwal Date: Mon, 2 Feb 2026 12:24:55 -0500 Subject: [PATCH 1/6] Add shebangs in bulk pre-testing --- language/gpt-oss-120b/setup.sh | 2 ++ language/llama2-70b/build.sh | 1 + language/llama2-70b/run_accuracy.sh | 2 ++ language/llama2-70b/run_offline.sh | 2 ++ language/llama2-70b/run_server.sh | 2 +- language/llama3.1-405b/build.sh | 2 ++ language/llama3.1-405b/run_accuracy.sh | 2 ++ language/llama3.1-405b/run_offline.sh | 2 ++ language/llama3.1-405b/run_server.sh | 2 +- language/llama3.1-8b/build.sh | 2 ++ language/llama3.1-8b/run_accuracy.sh | 2 ++ language/llama3.1-8b/run_offline.sh | 2 ++ language/llama3.1-8b/run_server.sh | 2 +- language/mixtral-8x7b/build.sh | 2 ++ language/mixtral-8x7b/run_accuracy.sh | 2 ++ language/mixtral-8x7b/run_offline.sh | 2 ++ language/mixtral-8x7b/run_server.sh | 2 +- tools/upscale_coco/upscale.sh | 2 ++ 18 files changed, 31 insertions(+), 4 deletions(-) diff --git a/language/gpt-oss-120b/setup.sh b/language/gpt-oss-120b/setup.sh index 23188a0cbd..41f42c266c 100755 --- a/language/gpt-oss-120b/setup.sh +++ b/language/gpt-oss-120b/setup.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + pip install -r requirements.txt git_dir=$(git rev-parse --show-toplevel) pip install $git_dir/loadgen \ No newline at end of file diff --git a/language/llama2-70b/build.sh b/language/llama2-70b/build.sh index 87afb992fa..999f8ea0fb 100644 --- a/language/llama2-70b/build.sh +++ b/language/llama2-70b/build.sh @@ -1,3 +1,4 @@ +#!/usr/bin/env bash set -e conda install pybind11==2.10.4 -c conda-forge -y diff --git a/language/llama2-70b/run_accuracy.sh b/language/llama2-70b/run_accuracy.sh index b4f7f8ad96..6e852a4bcd 100644 --- a/language/llama2-70b/run_accuracy.sh +++ b/language/llama2-70b/run_accuracy.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:-meta-llama/Llama-2-70b-chat-hf}" DATASET_PATH="${DATASET_PATH:-open-orca-val-set.pkl}" diff --git a/language/llama2-70b/run_offline.sh b/language/llama2-70b/run_offline.sh index 7153ea7cab..50a8e6a1ed 100644 --- a/language/llama2-70b/run_offline.sh +++ b/language/llama2-70b/run_offline.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:-meta-llama/Llama-2-70b-chat-hf}" DATASET_PATH="${DATASET_PATH:-open-orca-val-set.pkl}" diff --git a/language/llama2-70b/run_server.sh b/language/llama2-70b/run_server.sh index a71bf371e7..b19a6afcc6 100644 --- a/language/llama2-70b/run_server.sh +++ b/language/llama2-70b/run_server.sh @@ -1,4 +1,4 @@ - +#!/usr/bin/env bash CHECKPOINT_PATH="${CHECKPOINT_PATH:-meta-llama/Llama-2-70b-chat-hf}" DATASET_PATH="${DATASET_PATH:-open-orca-val-set.pkl}" diff --git a/language/llama3.1-405b/build.sh b/language/llama3.1-405b/build.sh index 87afb992fa..e7e7794cda 100644 --- a/language/llama3.1-405b/build.sh +++ b/language/llama3.1-405b/build.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + set -e conda install pybind11==2.10.4 -c conda-forge -y diff --git a/language/llama3.1-405b/run_accuracy.sh b/language/llama3.1-405b/run_accuracy.sh index 9a54d8f131..e4abe91210 100644 --- a/language/llama3.1-405b/run_accuracy.sh +++ b/language/llama3.1-405b/run_accuracy.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:Meta-Llama-3.1-405B-Instruct}" DATASET_PATH="${DATASET_PATH:mlperf_llama3.1_405b_dataset_8318.pkl}" diff --git a/language/llama3.1-405b/run_offline.sh b/language/llama3.1-405b/run_offline.sh index 6b3a56e012..e0b9f91c24 100644 --- a/language/llama3.1-405b/run_offline.sh +++ b/language/llama3.1-405b/run_offline.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:Meta-Llama-3.1-405B-Instruct}" DATASET_PATH="${DATASET_PATH:mlperf_llama3.1_405b_dataset_8318.pkl}" diff --git a/language/llama3.1-405b/run_server.sh b/language/llama3.1-405b/run_server.sh index 010a359de1..2197885ec8 100644 --- a/language/llama3.1-405b/run_server.sh +++ b/language/llama3.1-405b/run_server.sh @@ -1,4 +1,4 @@ - +#!/usr/bin/env bash CHECKPOINT_PATH="${CHECKPOINT_PATH:Meta-Llama-3.1-405B-Instruct}" DATASET_PATH="${DATASET_PATH:mlperf_llama3.1_405b_dataset_8318.pkl}" diff --git a/language/llama3.1-8b/build.sh b/language/llama3.1-8b/build.sh index 87afb992fa..e7e7794cda 100644 --- a/language/llama3.1-8b/build.sh +++ b/language/llama3.1-8b/build.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + set -e conda install pybind11==2.10.4 -c conda-forge -y diff --git a/language/llama3.1-8b/run_accuracy.sh b/language/llama3.1-8b/run_accuracy.sh index 52ec670339..e14e53544d 100644 --- a/language/llama3.1-8b/run_accuracy.sh +++ b/language/llama3.1-8b/run_accuracy.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:meta-llama/Meta-Llama-3.1-8B-Instruct}" DATASET_PATH="${DATASET_PATH:cnn_eval.json}" diff --git a/language/llama3.1-8b/run_offline.sh b/language/llama3.1-8b/run_offline.sh index 1d6bb271ad..7897d6ae69 100644 --- a/language/llama3.1-8b/run_offline.sh +++ b/language/llama3.1-8b/run_offline.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:meta-llama/Meta-Llama-3.1-8B-Instruct}" DATASET_PATH="${DATASET_PATH:cnn_eval.json}" diff --git a/language/llama3.1-8b/run_server.sh b/language/llama3.1-8b/run_server.sh index a870e9b37b..ba885aeba3 100644 --- a/language/llama3.1-8b/run_server.sh +++ b/language/llama3.1-8b/run_server.sh @@ -1,4 +1,4 @@ - +#!/usr/bin/env bash CHECKPOINT_PATH="${CHECKPOINT_PATH:meta-llama/Meta-Llama-3.1-8B-Instruct}" DATASET_PATH="${DATASET_PATH:cnn_eval.json}" diff --git a/language/mixtral-8x7b/build.sh b/language/mixtral-8x7b/build.sh index 87afb992fa..e7e7794cda 100644 --- a/language/mixtral-8x7b/build.sh +++ b/language/mixtral-8x7b/build.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + set -e conda install pybind11==2.10.4 -c conda-forge -y diff --git a/language/mixtral-8x7b/run_accuracy.sh b/language/mixtral-8x7b/run_accuracy.sh index 13da4e5daa..ae2c1e936b 100644 --- a/language/mixtral-8x7b/run_accuracy.sh +++ b/language/mixtral-8x7b/run_accuracy.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:mistralai/Mixtral-8x7B-Instruct-v0.1}" DATASET_PATH="${DATASET_PATH:dataset/2024_06_06_mixtral_15k_v4.pkl}" diff --git a/language/mixtral-8x7b/run_offline.sh b/language/mixtral-8x7b/run_offline.sh index 48afb50bd1..b32ed58ebc 100644 --- a/language/mixtral-8x7b/run_offline.sh +++ b/language/mixtral-8x7b/run_offline.sh @@ -1,3 +1,5 @@ +#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:mistralai/Mixtral-8x7B-Instruct-v0.1}" DATASET_PATH="${DATASET_PATH:dataset/2024_06_06_mixtral_15k_v4.pkl}" diff --git a/language/mixtral-8x7b/run_server.sh b/language/mixtral-8x7b/run_server.sh index ced7ea4432..f5ae85d082 100644 --- a/language/mixtral-8x7b/run_server.sh +++ b/language/mixtral-8x7b/run_server.sh @@ -1,4 +1,4 @@ - +#!/usr/bin/env bash CHECKPOINT_PATH="${CHECKPOINT_PATH:mistralai/Mixtral-8x7B-Instruct-v0.1}" DATASET_PATH="${DATASET_PATH:dataset/2024_06_06_mixtral_15k_v4.pkl}" diff --git a/tools/upscale_coco/upscale.sh b/tools/upscale_coco/upscale.sh index 1bcb7218ba..148e208297 100755 --- a/tools/upscale_coco/upscale.sh +++ b/tools/upscale_coco/upscale.sh @@ -1 +1,3 @@ +#!/usr/bin/env bash + python upscale_coco.py --inputs /scratch/ssd/data/coco/ --outputs ./coco700png --size 700 700 --format png From 3099c81739e0885618ec33b46542626bd956b8db Mon Sep 17 00:00:00 2001 From: Arav Agarwal Date: Tue, 3 Feb 2026 09:08:39 -0500 Subject: [PATCH 2/6] Update to exec mode torebase --- language/llama2-70b/build.sh | 0 language/llama2-70b/run_accuracy.sh | 0 language/llama2-70b/run_offline.sh | 0 language/llama2-70b/run_server.sh | 0 language/llama3.1-405b/build.sh | 0 language/llama3.1-405b/run_accuracy.sh | 0 language/llama3.1-405b/run_offline.sh | 0 language/llama3.1-405b/run_server.sh | 0 language/llama3.1-8b/build.sh | 0 language/llama3.1-8b/run_accuracy.sh | 0 language/llama3.1-8b/run_offline.sh | 0 language/llama3.1-8b/run_server.sh | 0 language/mixtral-8x7b/build.sh | 0 language/mixtral-8x7b/run_accuracy.sh | 0 language/mixtral-8x7b/run_offline.sh | 0 language/mixtral-8x7b/run_server.sh | 0 16 files changed, 0 insertions(+), 0 deletions(-) mode change 100644 => 100755 language/llama2-70b/build.sh mode change 100644 => 100755 language/llama2-70b/run_accuracy.sh mode change 100644 => 100755 language/llama2-70b/run_offline.sh mode change 100644 => 100755 language/llama2-70b/run_server.sh mode change 100644 => 100755 language/llama3.1-405b/build.sh mode change 100644 => 100755 language/llama3.1-405b/run_accuracy.sh mode change 100644 => 100755 language/llama3.1-405b/run_offline.sh mode change 100644 => 100755 language/llama3.1-405b/run_server.sh mode change 100644 => 100755 language/llama3.1-8b/build.sh mode change 100644 => 100755 language/llama3.1-8b/run_accuracy.sh mode change 100644 => 100755 language/llama3.1-8b/run_offline.sh mode change 100644 => 100755 language/llama3.1-8b/run_server.sh mode change 100644 => 100755 language/mixtral-8x7b/build.sh mode change 100644 => 100755 language/mixtral-8x7b/run_accuracy.sh mode change 100644 => 100755 language/mixtral-8x7b/run_offline.sh mode change 100644 => 100755 language/mixtral-8x7b/run_server.sh diff --git a/language/llama2-70b/build.sh b/language/llama2-70b/build.sh old mode 100644 new mode 100755 diff --git a/language/llama2-70b/run_accuracy.sh b/language/llama2-70b/run_accuracy.sh old mode 100644 new mode 100755 diff --git a/language/llama2-70b/run_offline.sh b/language/llama2-70b/run_offline.sh old mode 100644 new mode 100755 diff --git a/language/llama2-70b/run_server.sh b/language/llama2-70b/run_server.sh old mode 100644 new mode 100755 diff --git a/language/llama3.1-405b/build.sh b/language/llama3.1-405b/build.sh old mode 100644 new mode 100755 diff --git a/language/llama3.1-405b/run_accuracy.sh b/language/llama3.1-405b/run_accuracy.sh old mode 100644 new mode 100755 diff --git a/language/llama3.1-405b/run_offline.sh b/language/llama3.1-405b/run_offline.sh old mode 100644 new mode 100755 diff --git a/language/llama3.1-405b/run_server.sh b/language/llama3.1-405b/run_server.sh old mode 100644 new mode 100755 diff --git a/language/llama3.1-8b/build.sh b/language/llama3.1-8b/build.sh old mode 100644 new mode 100755 diff --git a/language/llama3.1-8b/run_accuracy.sh b/language/llama3.1-8b/run_accuracy.sh old mode 100644 new mode 100755 diff --git a/language/llama3.1-8b/run_offline.sh b/language/llama3.1-8b/run_offline.sh old mode 100644 new mode 100755 diff --git a/language/llama3.1-8b/run_server.sh b/language/llama3.1-8b/run_server.sh old mode 100644 new mode 100755 diff --git a/language/mixtral-8x7b/build.sh b/language/mixtral-8x7b/build.sh old mode 100644 new mode 100755 diff --git a/language/mixtral-8x7b/run_accuracy.sh b/language/mixtral-8x7b/run_accuracy.sh old mode 100644 new mode 100755 diff --git a/language/mixtral-8x7b/run_offline.sh b/language/mixtral-8x7b/run_offline.sh old mode 100644 new mode 100755 diff --git a/language/mixtral-8x7b/run_server.sh b/language/mixtral-8x7b/run_server.sh old mode 100644 new mode 100755 From 134a6528cdc23ddb2c48138e76b8f4ee38169bb9 Mon Sep 17 00:00:00 2001 From: Arav Agarwal Date: Mon, 9 Feb 2026 07:00:38 -0500 Subject: [PATCH 3/6] Remove use of eval from torch quantization system --- .../tools/calibrate_torchvision_model.py | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/vision/classification_and_detection/tools/calibrate_torchvision_model.py b/vision/classification_and_detection/tools/calibrate_torchvision_model.py index 3b002003ab..875d26c388 100644 --- a/vision/classification_and_detection/tools/calibrate_torchvision_model.py +++ b/vision/classification_and_detection/tools/calibrate_torchvision_model.py @@ -7,7 +7,7 @@ from torch.utils.data import DataLoader, Dataset import torchvision.transforms as transforms -from torchvision.models.quantization import * +import torchvision.models.quantization as torchvision_quantization_models class CalibrationDataset(Dataset): @@ -73,7 +73,11 @@ def main(): ) dataloader = DataLoader(dataset, batch_size=1) - model = eval(args.model)(pretrained=True, progress=True, quantize=False) + if not hasattr(torchvision_quantization_models, args.model): + raise ValueError(f"Model {args.model} not found in torchvision quantization models") + + + model = getattr(torchvision_quantization_models, args.model)(pretrained=True, progress=True, quantize=False) quantize_model(model, dataloader) print(model) From 9f64a04fc2f7a5f0fbd54086187a953ac88051f8 Mon Sep 17 00:00:00 2001 From: Arav Agarwal Date: Mon, 9 Feb 2026 08:08:01 -0500 Subject: [PATCH 4/6] Revert "Add shebangs in bulk pre-testing" This reverts commit 0ee8fb9560b9412ab635a213cf17cff9b883a8aa. --- language/gpt-oss-120b/setup.sh | 2 -- language/llama2-70b/build.sh | 1 - language/llama2-70b/run_accuracy.sh | 2 -- language/llama2-70b/run_offline.sh | 2 -- language/llama2-70b/run_server.sh | 2 +- language/llama3.1-405b/build.sh | 2 -- language/llama3.1-405b/run_accuracy.sh | 2 -- language/llama3.1-405b/run_offline.sh | 2 -- language/llama3.1-405b/run_server.sh | 2 +- language/llama3.1-8b/build.sh | 2 -- language/llama3.1-8b/run_accuracy.sh | 2 -- language/llama3.1-8b/run_offline.sh | 2 -- language/llama3.1-8b/run_server.sh | 2 +- language/mixtral-8x7b/build.sh | 2 -- language/mixtral-8x7b/run_accuracy.sh | 2 -- language/mixtral-8x7b/run_offline.sh | 2 -- language/mixtral-8x7b/run_server.sh | 2 +- tools/upscale_coco/upscale.sh | 2 -- 18 files changed, 4 insertions(+), 31 deletions(-) diff --git a/language/gpt-oss-120b/setup.sh b/language/gpt-oss-120b/setup.sh index 41f42c266c..23188a0cbd 100755 --- a/language/gpt-oss-120b/setup.sh +++ b/language/gpt-oss-120b/setup.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - pip install -r requirements.txt git_dir=$(git rev-parse --show-toplevel) pip install $git_dir/loadgen \ No newline at end of file diff --git a/language/llama2-70b/build.sh b/language/llama2-70b/build.sh index 999f8ea0fb..87afb992fa 100755 --- a/language/llama2-70b/build.sh +++ b/language/llama2-70b/build.sh @@ -1,4 +1,3 @@ -#!/usr/bin/env bash set -e conda install pybind11==2.10.4 -c conda-forge -y diff --git a/language/llama2-70b/run_accuracy.sh b/language/llama2-70b/run_accuracy.sh index 6e852a4bcd..b4f7f8ad96 100755 --- a/language/llama2-70b/run_accuracy.sh +++ b/language/llama2-70b/run_accuracy.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - CHECKPOINT_PATH="${CHECKPOINT_PATH:-meta-llama/Llama-2-70b-chat-hf}" DATASET_PATH="${DATASET_PATH:-open-orca-val-set.pkl}" diff --git a/language/llama2-70b/run_offline.sh b/language/llama2-70b/run_offline.sh index 50a8e6a1ed..7153ea7cab 100755 --- a/language/llama2-70b/run_offline.sh +++ b/language/llama2-70b/run_offline.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - CHECKPOINT_PATH="${CHECKPOINT_PATH:-meta-llama/Llama-2-70b-chat-hf}" DATASET_PATH="${DATASET_PATH:-open-orca-val-set.pkl}" diff --git a/language/llama2-70b/run_server.sh b/language/llama2-70b/run_server.sh index b19a6afcc6..a71bf371e7 100755 --- a/language/llama2-70b/run_server.sh +++ b/language/llama2-70b/run_server.sh @@ -1,4 +1,4 @@ -#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:-meta-llama/Llama-2-70b-chat-hf}" DATASET_PATH="${DATASET_PATH:-open-orca-val-set.pkl}" diff --git a/language/llama3.1-405b/build.sh b/language/llama3.1-405b/build.sh index e7e7794cda..87afb992fa 100755 --- a/language/llama3.1-405b/build.sh +++ b/language/llama3.1-405b/build.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - set -e conda install pybind11==2.10.4 -c conda-forge -y diff --git a/language/llama3.1-405b/run_accuracy.sh b/language/llama3.1-405b/run_accuracy.sh index e4abe91210..9a54d8f131 100755 --- a/language/llama3.1-405b/run_accuracy.sh +++ b/language/llama3.1-405b/run_accuracy.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - CHECKPOINT_PATH="${CHECKPOINT_PATH:Meta-Llama-3.1-405B-Instruct}" DATASET_PATH="${DATASET_PATH:mlperf_llama3.1_405b_dataset_8318.pkl}" diff --git a/language/llama3.1-405b/run_offline.sh b/language/llama3.1-405b/run_offline.sh index e0b9f91c24..6b3a56e012 100755 --- a/language/llama3.1-405b/run_offline.sh +++ b/language/llama3.1-405b/run_offline.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - CHECKPOINT_PATH="${CHECKPOINT_PATH:Meta-Llama-3.1-405B-Instruct}" DATASET_PATH="${DATASET_PATH:mlperf_llama3.1_405b_dataset_8318.pkl}" diff --git a/language/llama3.1-405b/run_server.sh b/language/llama3.1-405b/run_server.sh index 2197885ec8..010a359de1 100755 --- a/language/llama3.1-405b/run_server.sh +++ b/language/llama3.1-405b/run_server.sh @@ -1,4 +1,4 @@ -#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:Meta-Llama-3.1-405B-Instruct}" DATASET_PATH="${DATASET_PATH:mlperf_llama3.1_405b_dataset_8318.pkl}" diff --git a/language/llama3.1-8b/build.sh b/language/llama3.1-8b/build.sh index e7e7794cda..87afb992fa 100755 --- a/language/llama3.1-8b/build.sh +++ b/language/llama3.1-8b/build.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - set -e conda install pybind11==2.10.4 -c conda-forge -y diff --git a/language/llama3.1-8b/run_accuracy.sh b/language/llama3.1-8b/run_accuracy.sh index e14e53544d..52ec670339 100755 --- a/language/llama3.1-8b/run_accuracy.sh +++ b/language/llama3.1-8b/run_accuracy.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - CHECKPOINT_PATH="${CHECKPOINT_PATH:meta-llama/Meta-Llama-3.1-8B-Instruct}" DATASET_PATH="${DATASET_PATH:cnn_eval.json}" diff --git a/language/llama3.1-8b/run_offline.sh b/language/llama3.1-8b/run_offline.sh index 7897d6ae69..1d6bb271ad 100755 --- a/language/llama3.1-8b/run_offline.sh +++ b/language/llama3.1-8b/run_offline.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - CHECKPOINT_PATH="${CHECKPOINT_PATH:meta-llama/Meta-Llama-3.1-8B-Instruct}" DATASET_PATH="${DATASET_PATH:cnn_eval.json}" diff --git a/language/llama3.1-8b/run_server.sh b/language/llama3.1-8b/run_server.sh index ba885aeba3..a870e9b37b 100755 --- a/language/llama3.1-8b/run_server.sh +++ b/language/llama3.1-8b/run_server.sh @@ -1,4 +1,4 @@ -#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:meta-llama/Meta-Llama-3.1-8B-Instruct}" DATASET_PATH="${DATASET_PATH:cnn_eval.json}" diff --git a/language/mixtral-8x7b/build.sh b/language/mixtral-8x7b/build.sh index e7e7794cda..87afb992fa 100755 --- a/language/mixtral-8x7b/build.sh +++ b/language/mixtral-8x7b/build.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - set -e conda install pybind11==2.10.4 -c conda-forge -y diff --git a/language/mixtral-8x7b/run_accuracy.sh b/language/mixtral-8x7b/run_accuracy.sh index ae2c1e936b..13da4e5daa 100755 --- a/language/mixtral-8x7b/run_accuracy.sh +++ b/language/mixtral-8x7b/run_accuracy.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - CHECKPOINT_PATH="${CHECKPOINT_PATH:mistralai/Mixtral-8x7B-Instruct-v0.1}" DATASET_PATH="${DATASET_PATH:dataset/2024_06_06_mixtral_15k_v4.pkl}" diff --git a/language/mixtral-8x7b/run_offline.sh b/language/mixtral-8x7b/run_offline.sh index b32ed58ebc..48afb50bd1 100755 --- a/language/mixtral-8x7b/run_offline.sh +++ b/language/mixtral-8x7b/run_offline.sh @@ -1,5 +1,3 @@ -#!/usr/bin/env bash - CHECKPOINT_PATH="${CHECKPOINT_PATH:mistralai/Mixtral-8x7B-Instruct-v0.1}" DATASET_PATH="${DATASET_PATH:dataset/2024_06_06_mixtral_15k_v4.pkl}" diff --git a/language/mixtral-8x7b/run_server.sh b/language/mixtral-8x7b/run_server.sh index f5ae85d082..ced7ea4432 100755 --- a/language/mixtral-8x7b/run_server.sh +++ b/language/mixtral-8x7b/run_server.sh @@ -1,4 +1,4 @@ -#!/usr/bin/env bash + CHECKPOINT_PATH="${CHECKPOINT_PATH:mistralai/Mixtral-8x7B-Instruct-v0.1}" DATASET_PATH="${DATASET_PATH:dataset/2024_06_06_mixtral_15k_v4.pkl}" diff --git a/tools/upscale_coco/upscale.sh b/tools/upscale_coco/upscale.sh index 148e208297..1bcb7218ba 100755 --- a/tools/upscale_coco/upscale.sh +++ b/tools/upscale_coco/upscale.sh @@ -1,3 +1 @@ -#!/usr/bin/env bash - python upscale_coco.py --inputs /scratch/ssd/data/coco/ --outputs ./coco700png --size 700 700 --format png From 0d77cebe59787a1fd70df59335832c683a510ce4 Mon Sep 17 00:00:00 2001 From: Arav Agarwal Date: Mon, 9 Feb 2026 08:08:56 -0500 Subject: [PATCH 5/6] Revert "Update to exec mode" This reverts commit 3099c81739e0885618ec33b46542626bd956b8db. --- language/llama2-70b/build.sh | 0 language/llama2-70b/run_accuracy.sh | 0 language/llama2-70b/run_offline.sh | 0 language/llama2-70b/run_server.sh | 0 language/llama3.1-405b/build.sh | 0 language/llama3.1-405b/run_accuracy.sh | 0 language/llama3.1-405b/run_offline.sh | 0 language/llama3.1-405b/run_server.sh | 0 language/llama3.1-8b/build.sh | 0 language/llama3.1-8b/run_accuracy.sh | 0 language/llama3.1-8b/run_offline.sh | 0 language/llama3.1-8b/run_server.sh | 0 language/mixtral-8x7b/build.sh | 0 language/mixtral-8x7b/run_accuracy.sh | 0 language/mixtral-8x7b/run_offline.sh | 0 language/mixtral-8x7b/run_server.sh | 0 16 files changed, 0 insertions(+), 0 deletions(-) mode change 100755 => 100644 language/llama2-70b/build.sh mode change 100755 => 100644 language/llama2-70b/run_accuracy.sh mode change 100755 => 100644 language/llama2-70b/run_offline.sh mode change 100755 => 100644 language/llama2-70b/run_server.sh mode change 100755 => 100644 language/llama3.1-405b/build.sh mode change 100755 => 100644 language/llama3.1-405b/run_accuracy.sh mode change 100755 => 100644 language/llama3.1-405b/run_offline.sh mode change 100755 => 100644 language/llama3.1-405b/run_server.sh mode change 100755 => 100644 language/llama3.1-8b/build.sh mode change 100755 => 100644 language/llama3.1-8b/run_accuracy.sh mode change 100755 => 100644 language/llama3.1-8b/run_offline.sh mode change 100755 => 100644 language/llama3.1-8b/run_server.sh mode change 100755 => 100644 language/mixtral-8x7b/build.sh mode change 100755 => 100644 language/mixtral-8x7b/run_accuracy.sh mode change 100755 => 100644 language/mixtral-8x7b/run_offline.sh mode change 100755 => 100644 language/mixtral-8x7b/run_server.sh diff --git a/language/llama2-70b/build.sh b/language/llama2-70b/build.sh old mode 100755 new mode 100644 diff --git a/language/llama2-70b/run_accuracy.sh b/language/llama2-70b/run_accuracy.sh old mode 100755 new mode 100644 diff --git a/language/llama2-70b/run_offline.sh b/language/llama2-70b/run_offline.sh old mode 100755 new mode 100644 diff --git a/language/llama2-70b/run_server.sh b/language/llama2-70b/run_server.sh old mode 100755 new mode 100644 diff --git a/language/llama3.1-405b/build.sh b/language/llama3.1-405b/build.sh old mode 100755 new mode 100644 diff --git a/language/llama3.1-405b/run_accuracy.sh b/language/llama3.1-405b/run_accuracy.sh old mode 100755 new mode 100644 diff --git a/language/llama3.1-405b/run_offline.sh b/language/llama3.1-405b/run_offline.sh old mode 100755 new mode 100644 diff --git a/language/llama3.1-405b/run_server.sh b/language/llama3.1-405b/run_server.sh old mode 100755 new mode 100644 diff --git a/language/llama3.1-8b/build.sh b/language/llama3.1-8b/build.sh old mode 100755 new mode 100644 diff --git a/language/llama3.1-8b/run_accuracy.sh b/language/llama3.1-8b/run_accuracy.sh old mode 100755 new mode 100644 diff --git a/language/llama3.1-8b/run_offline.sh b/language/llama3.1-8b/run_offline.sh old mode 100755 new mode 100644 diff --git a/language/llama3.1-8b/run_server.sh b/language/llama3.1-8b/run_server.sh old mode 100755 new mode 100644 diff --git a/language/mixtral-8x7b/build.sh b/language/mixtral-8x7b/build.sh old mode 100755 new mode 100644 diff --git a/language/mixtral-8x7b/run_accuracy.sh b/language/mixtral-8x7b/run_accuracy.sh old mode 100755 new mode 100644 diff --git a/language/mixtral-8x7b/run_offline.sh b/language/mixtral-8x7b/run_offline.sh old mode 100755 new mode 100644 diff --git a/language/mixtral-8x7b/run_server.sh b/language/mixtral-8x7b/run_server.sh old mode 100755 new mode 100644 From 5fe3dcc517bd828030c537eb2f10d95c964af834 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 9 Feb 2026 13:48:18 +0000 Subject: [PATCH 6/6] [Automated Commit] Format Codebase --- text_to_video/wan-2.2-t2v-a14b/run_mlperf.py | 28 ++++++++++++++------ 1 file changed, 20 insertions(+), 8 deletions(-) diff --git a/text_to_video/wan-2.2-t2v-a14b/run_mlperf.py b/text_to_video/wan-2.2-t2v-a14b/run_mlperf.py index 147624b340..ab73c25966 100644 --- a/text_to_video/wan-2.2-t2v-a14b/run_mlperf.py +++ b/text_to_video/wan-2.2-t2v-a14b/run_mlperf.py @@ -46,7 +46,8 @@ def load_prompts(dataset_path): class Model: - def __init__(self, model_path, device, config, prompts, fixed_latent=None, rank=0): + def __init__(self, model_path, device, config, + prompts, fixed_latent=None, rank=0): self.device = device self.rank = rank self.height = config["height"] @@ -106,7 +107,8 @@ def flush_queries(self): class DebugModel: - def __init__(self, model_path, device, config, prompts, fixed_latent=None, rank=0): + def __init__(self, model_path, device, config, + prompts, fixed_latent=None, rank=0): self.prompts = prompts def issue_queries(self, query_samples): @@ -186,7 +188,8 @@ def get_args(): parser.add_argument( "--scenario", default="SingleStream", - help="mlperf benchmark scenario, one of " + str(list(SCENARIO_MAP.keys())), + help="mlperf benchmark scenario, one of " + + str(list(SCENARIO_MAP.keys())), ) parser.add_argument( "--user_conf", @@ -202,7 +205,10 @@ def get_args(): help="performance sample count", default=5000, ) - parser.add_argument("--accuracy", action="store_true", help="enable accuracy pass") + parser.add_argument( + "--accuracy", + action="store_true", + help="enable accuracy pass") # Dont overwrite these for official submission parser.add_argument("--count", type=int, help="dataset items to use") parser.add_argument("--time", type=int, help="time to scan in seconds") @@ -271,7 +277,10 @@ def run_mlperf(args, config): audit_config = os.path.abspath(args.audit_conf) if os.path.exists(audit_config): - settings.FromConfig(audit_config, "wan-2.2-t2v-a14b", args.scenario) + settings.FromConfig( + audit_config, + "wan-2.2-t2v-a14b", + args.scenario) settings.scenario = SCENARIO_MAP[args.scenario] settings.mode = lg.TestMode.PerformanceOnly @@ -297,8 +306,10 @@ def run_mlperf(args, config): if args.samples_per_query: settings.multi_stream_samples_per_query = args.samples_per_query if args.max_latency: - settings.server_target_latency_ns = int(args.max_latency * NANO_SEC) - settings.multi_stream_expected_latency_ns = int(args.max_latency * NANO_SEC) + settings.server_target_latency_ns = int( + args.max_latency * NANO_SEC) + settings.multi_stream_expected_latency_ns = int( + args.max_latency * NANO_SEC) performance_sample_count = ( args.performance_sample_count @@ -311,7 +322,8 @@ def run_mlperf(args, config): count, performance_sample_count, load_query_samples, unload_query_samples ) - lg.StartTestWithLogSettings(sut, qsl, settings, log_settings, audit_config) + lg.StartTestWithLogSettings( + sut, qsl, settings, log_settings, audit_config) lg.DestroyQSL(qsl) lg.DestroySUT(sut)