Skip to content

drai-inn/llm-nesi-example

Repository files navigation

Example running an LLM on NeSI

VM configuration

Note: much of this can be automated...

See also: https://docs.litellm.ai/docs/tutorials/openweb_ui

Prerequisites:

Create a .env file alongside the docker-compose.yaml (fill in the values):

# .env file contents
LITELLM_DB_PASSWORD=
LITELLM_MASTER_KEY=sk-
LITELLM_SALT_KEY=sk-
LITELLM_UI_PASSWORD=
SWAG_DOMAIN=
OPEN_WEBUI_URL=
VLLM_API_KEY=
FLOWISE_USERNAME=
FLOWISE_PASSWORD=
# virtual key created in the litellm ui manually:
OPENAI_API_KEY=sk-

Run it: sudo docker compose up -d

After running, need to connect to LiteLLM (port 4000) user interface and create a virtual key, then use that in the .env file above.

Also after running the above, need to insert the DuckDNS API key into swag-data/dns-conf/duckdns.ini.

HPC side configuration

Pull the vllm docker image using apptainer (gptoss tag required as of 2025-08-12)...

Non-interactive SSH to the VM must be configured...

Configure/edit the example Slurm script and submit to Slurm.

Ports, API keys, model names, etc. should match between the Slurm script and .env file.

About

Running an LLM on NeSI for remote interactive inference

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors