forked from thammegowda/wmt25-model-compression
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathDockerfile
More file actions
executable file
·35 lines (28 loc) · 1.49 KB
/
Dockerfile
File metadata and controls
executable file
·35 lines (28 loc) · 1.49 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
FROM nvidia/cuda:12.6.0-cudnn-devel-ubuntu22.04
LABEL description="Dockerfile for WMT25 Model Compression Shared Task"
LABEL maintainer="WMT25 Model Compression Task Organizers"
LABEL version="1.0"
LABEL date="2025-05-15"
ARG DEBIAN_FRONTEND=noninteractive
# !!! Follow the instructions in README.md before building this image !!!
# Install default packages
RUN apt update && apt upgrade --fix-missing -y
RUN apt install -y git python3 python3-pip emacs-nox vim wget curl ncdu tmux htop tree && \
apt clean && rm -rf /var/lib/apt/lists/*
RUN python3 -m pip install --no-cache-dir --upgrade pip
# install torch built against the CUDA version of docker image; e.g. 12.8
# RUN python3 -m pip install --no-cache-dir torch==2.7.0 --index-url https://download.pytorch.org/whl/cu128
WORKDIR /work/wmt25-model-compression
COPY requirements.txt pyproject.toml README.md ./
COPY modelzip/ modelzip/
RUN ls -lh && python3 -m pip install --no-cache-dir -e ./
RUN python3 -m modelzip.setup -h
##==============================================
# Copy model files with execution scripts
# Note: these models are for demonstration purposes only
# Do not include these in the submission image, include your compressed model(s) instead
#COPY workdir/models/aya-expanse-8b-bnb-8bit /model/bnb-8bit
#COPY workdir/models/aya-expanse-8b-bnb-4bit /model/bnb-4bit
#
#RUN bash /model/bnb-8bit/run.sh ces-deu 1 <<< "This is a test with the 8-bit model."
#RUN bash /model/bnb-4bit/run.sh ces-deu 1 <<< "This is a test with the 4-bit model."