Skip to content

Latest commit

 

History

History
34 lines (25 loc) · 1.1 KB

File metadata and controls

34 lines (25 loc) · 1.1 KB

If your network is restricted, please download the dependencies manually from the following links:

# Download the LIBERO submodule
git clone https://github.com/Lifelong-Robot-Learning/LIBERO.git third_party/LIBERO

# Download flash-attention wheel
wget https://github.com/Dao-AILab/flash-attention/releases/download/v2.6.0.post1/flash_attn-2.6.0.post1+cu122torch2.4cxx11abiFALSE-cp310-cp310-linux_x86_64.whl -P third_party/

# Download dlimp package
git clone https://github.com/moojink/dlimp.git third_party/dlimp

Then, follow these steps to install the dependencies:

# 1) Set up the environment
uv venv --seed -p 3.10
source .venv/bin/activate

# 2) Install dependencies
uv pip install -e train/verl/".[gpu]"
uv pip install -e third_party/flash-attn-2.6.0.post1+cu122torch2.4cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
uv pip install -e train/verl/".[vllm]"
uv pip install -r train/verl/requirements.txt

# 3) Install vla-adapter
uv pip install -e third_party/dlimp
uv pip install -e train/verl/vla-adapter/openvla-oft

# 4) Install LIBERO requirements
uv pip install -e third_party/LIBERO