diff --git a/README.md b/README.md index 8819787..4cfaa0d 100644 --- a/README.md +++ b/README.md @@ -374,11 +374,11 @@ conda create --name perception_models python=3.12 conda activate perception_models # Install PyTorch -pip install torch==2.5.1 torchvision==0.20.1 torchaudio==2.5.1 xformers --index-url https://download.pytorch.org/whl/cu124 +pip install torch torchvision torchaudio xformers --index-url https://download.pytorch.org/whl/cu124 # We use torchcodec for decoding videos into PyTorch tensors conda install ffmpeg -c conda-forge -pip install torchcodec==0.1 --index-url=https://download.pytorch.org/whl/cu124 +pip install torchcodec --index-url=https://download.pytorch.org/whl/cu124 pip install -e . ``` diff --git a/requirements.txt b/requirements.txt index 2ac27f8..60add7a 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,33 +1,34 @@ -numpy==2.1.2 -omegaconf==2.3.0 -msgspec==0.19.0 -rouge-score==0.1.2 -sacrebleu==2.5.1 -sentencepiece==0.2.0 -tiktoken==0.9.0 -blobfile==3.0.0 -wandb==0.19.8 -viztracer==1.0.3 -lm-eval==0.4.8 -scipy==1.15.2 -pynvml==12.0.0 -orjson==3.10.15 -einops==0.8.1 -pillow==11.0.0 -pyahocorasick==2.1.0 -iopath==0.1.10 -torchdata==0.11.0 +numpy +omegaconf +msgspec +rouge-score +sacrebleu +sentencepiece +tiktoken +blobfile +wandb +viztracer +lm-eval +scipy +pynvml +orjson +einops +pillow +pyahocorasick +iopath +torchdata torchcodec -timm==1.0.15 -decord==0.6.0 -opencv-python==4.11.0.86 -pycocoevalcap==1.2 -scikit-learn==1.6.1 -scipy==1.15.2 -sentencepiece==0.2.0 -tokenizers==0.21.1 -webdataset==0.2.111 +timm +decord +opencv-python +pycocoevalcap +scikit-learn +scipy +sentencepiece +tokenizers +webdataset fsspec datatrove ftfy transformers>=4.48.0 +xformers