Skip to content

Update rl/unified/README#2404

Open
daniellepintz wants to merge 2 commits intomainfrom
dp/readme
Open

Update rl/unified/README#2404
daniellepintz wants to merge 2 commits intomainfrom
dp/readme

Conversation

@daniellepintz
Copy link
Contributor

  • add steps to create env
  • add note about setting LD_PRELOAD
  • add step to install torchtitan in editable mode

@daniellepintz daniellepintz requested a review from wwwjn February 20, 2026 11:59
@meta-cla meta-cla bot added the CLA Signed This label is managed by the Meta Open Source bot. label Feb 20, 2026

**NOTE:** The pre-built vLLM wheels are only compatible with CUDA 12.8, though they should work with most older CUDA versions. Alternatively, you can install the corresponding vLLM pre-built wheels directly from https://download.pytorch.org/whl/nightly/cu128, for example: `uv pip install vllm-1.0.0.dev20260219+cu130-<suffix>.whl`. Ensure the build version number (e.g., `dev20260219`) matches your PyTorch nightly installation.

**NOTE:** If you run into a vLLM import error, you may have to set the LD_PRELOAD environment variable to point to libcublas.so.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks, is this a common issue? Can you be more specific what is "vLLM import error" here? Maybe not all import issue are related to can not find cublas

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The error was undefined symbol: cublasGemmEx. Yeah I can add this to the README; I am not sure how common this is, but I ran into it unfortunately :/ Lmk if you think it makes sense to include this note in the README

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what was the command you needed to use on your machine to fix that?

Copy link
Contributor

@allenwang28 allenwang28 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

good fixes!


0. Create environment with uv:
```bash
uv venv --python 3.12 titan-rl
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

why don't we set these in a pyproject.toml? like an [rl] extra tag. then we can just uv sync --extra rl

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

sgtm! @wwwjn @tianyu-l does that make sense to you if we put this in the main torchtitan pyproject.toml?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

torchtitan haven't extensively move to uv for now, this setup step here seems easy and we can just limit our change to RL folder for now


**NOTE:** The pre-built vLLM wheels are only compatible with CUDA 12.8, though they should work with most older CUDA versions. Alternatively, you can install the corresponding vLLM pre-built wheels directly from https://download.pytorch.org/whl/nightly/cu128, for example: `uv pip install vllm-1.0.0.dev20260219+cu130-<suffix>.whl`. Ensure the build version number (e.g., `dev20260219`) matches your PyTorch nightly installation.

**NOTE:** If you run into a vLLM import error, you may have to set the LD_PRELOAD environment variable to point to libcublas.so.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what was the command you needed to use on your machine to fix that?

Copy link
Contributor

@wwwjn wwwjn left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please address the comments before merge, thank you


0. Create environment with uv:
```bash
uv venv --python 3.12 titan-rl
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

torchtitan haven't extensively move to uv for now, this setup step here seems easy and we can just limit our change to RL folder for now


**NOTE:** The pre-built vLLM wheels are only compatible with CUDA 12.8, though they should work with most older CUDA versions. Alternatively, you can install the corresponding vLLM pre-built wheels directly from https://download.pytorch.org/whl/nightly/cu128, for example: `uv pip install vllm-1.0.0.dev20260219+cu130-<suffix>.whl`. Ensure the build version number (e.g., `dev20260219`) matches your PyTorch nightly installation.

**NOTE:** If you run into a `undefined symbol: cublasGemmEx` error when importing vLLM, you may have to set the LD_PRELOAD environment variable to point to libcublas.so.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks like this is very related to local cuda setup issue. Can we remove this note to make README simplify?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/8gpu CLA Signed This label is managed by the Meta Open Source bot.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants