diff --git a/.pin/constraints-cuda-torch.txt b/.pin/constraints-cuda-torch.txt index d6e5ae440..eb53eeab9 100644 --- a/.pin/constraints-cuda-torch.txt +++ b/.pin/constraints-cuda-torch.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=.pin/constraints-cuda-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in +# pip-compile --output-file=.pin/constraints-cuda-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in # --extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://download.pytorch.org/whl/cu121 @@ -28,6 +28,8 @@ aiohttp==3.9.5 # fsspec aiosignal==1.3.1 # via aiohttp +annotated-types==0.7.0 + # via pydantic antlr4-python3-runtime==4.9.3 # via omegaconf asttokens==2.4.1 @@ -52,16 +54,14 @@ codefind==0.1.6 # via ptera contextlib2==21.6.0 # via ml-collections -datasets==2.19.2 +datasets==2.14.7 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/llama/requirements.in # evaluate deepspeed==0.14.2 - # via - # -r benchmarks/accelerate_opt/requirements.in - # -r benchmarks/rwkv/requirements.in -dill==0.3.8 + # via -r benchmarks/accelerate_opt/requirements.in +dill==0.3.7 # via # datasets # evaluate @@ -72,7 +72,7 @@ dm-tree==0.1.8 # via dm-env docker==7.1.0 # via torchx -docstring-parser==0.8.1 +docstring-parser==0.16 # via torchx etils[epath,epy]==1.9.1 # via @@ -90,7 +90,6 @@ fbgemm-gpu==0.7.0+cu121 # via torchrec filelock==3.14.0 # via - # datasets # huggingface-hub # torch # torchx @@ -110,13 +109,12 @@ frozenlist==1.4.1 # via # aiohttp # aiosignal -fsspec[http]==2024.3.1 +fsspec[http]==2023.1.0 # via # datasets # etils # evaluate # huggingface-hub - # pytorch-lightning # torch # torchx future==1.0.0 @@ -139,7 +137,7 @@ gym-notices==0.0.8 # via gym hjson==3.1.0 # via deepspeed -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -r benchmarks/timm/requirements.in # accelerate @@ -192,9 +190,7 @@ jinja2==3.1.4 joblib==1.4.2 # via scikit-learn lightning-utilities==0.11.2 - # via - # pytorch-lightning - # torchmetrics + # via torchmetrics markdown==3.6 # via tensorboard markdown-it-py==3.0.0 @@ -228,7 +224,7 @@ multidict==6.0.5 # via # aiohttp # yarl -multiprocess==0.70.16 +multiprocess==0.70.15 # via # datasets # evaluate @@ -239,13 +235,10 @@ nest-asyncio==1.6.0 networkx==3.3 # via torch ninja==1.11.1.1 - # via - # -r benchmarks/rwkv/requirements.in - # deepspeed + # via deepspeed numpy==1.26.4 # via # -r benchmarks/dlrm/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # accelerate @@ -271,7 +264,6 @@ numpy==1.26.4 # orbax-checkpoint # pandas # pyarrow - # pytorch-lightning # scikit-learn # scipy # tensorboard @@ -355,7 +347,6 @@ packaging==24.1 # evaluate # huggingface-hub # lightning-utilities - # pytorch-lightning # tensorboardx # torchmetrics # transformers @@ -386,10 +377,10 @@ pyarrow==16.1.0 # via datasets pyarrow-hotfix==0.6 # via datasets -pydantic==1.10.15 - # via - # -r benchmarks/rwkv/requirements.in - # deepspeed +pydantic==2.7.3 + # via deepspeed +pydantic-core==2.18.4 + # via pydantic pydot==2.0.0 # via -r benchmarks/dlrm/requirements.in pygments==2.18.0 @@ -408,8 +399,6 @@ python-dateutil==2.9.0.post0 # via pandas pytinyrenderer==0.0.14 # via brax -pytorch-lightning==1.9.5 - # via -r benchmarks/rwkv/requirements.in pytz==2024.1 # via pandas pyyaml==6.0.1 @@ -422,7 +411,6 @@ pyyaml==6.0.1 # ml-collections # omegaconf # orbax-checkpoint - # pytorch-lightning # torchx # transformers reactivex==4.0.4 @@ -434,6 +422,7 @@ requests==2.32.3 # datasets # docker # evaluate + # fsspec # huggingface-hub # transformers rich==13.7.1 @@ -475,7 +464,7 @@ tensorboard-data-server==0.7.2 # via tensorboard tensorboardx==2.6.2.2 # via brax -tensorstore==0.1.60 +tensorstore==0.1.61 # via # flax # orbax-checkpoint @@ -483,7 +472,7 @@ termcolor==2.4.0 # via fire threadpoolctl==3.5.0 # via scikit-learn -tokenizers==0.19.1 +tokenizers==0.15.2 # via transformers toolz==0.12.1 # via chex @@ -495,7 +484,6 @@ torch==2.3.1+cu121 # -r benchmarks/flops/requirements.in # -r benchmarks/huggingface/requirements.in # -r benchmarks/llama/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # -r benchmarks/timm/requirements.in @@ -504,7 +492,6 @@ torch==2.3.1+cu121 # accelerate # deepspeed # fairscale - # pytorch-lightning # torchaudio # torchcompat # torchmetrics @@ -519,9 +506,7 @@ torchcompat==1.0.2 # -r benchmarks/torchvision/requirements.in # -r benchmarks/torchvision_ddp/requirements.in torchmetrics==1.0.3 - # via - # pytorch-lightning - # torchrec + # via torchrec torchrec==0.7.0+cu121 # via -r benchmarks/dlrm/requirements.in torchvision==0.18.1+cu121 @@ -535,7 +520,7 @@ torchvision==0.18.1+cu121 # -r benchmarks/torchvision_ddp/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in -torchx==0.5.0 +torchx==0.6.0 # via -r benchmarks/dlrm/requirements.in tqdm==4.66.4 # via @@ -548,10 +533,9 @@ tqdm==4.66.4 # deepspeed # evaluate # huggingface-hub - # pytorch-lightning # torchrec # transformers -transformers==4.41.2 +transformers==4.35.2 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/huggingface/requirements.in @@ -572,8 +556,8 @@ typing-extensions==4.12.2 # lightning-utilities # orbax-checkpoint # pydantic + # pydantic-core # pyre-extensions - # pytorch-lightning # reactivex # torch # typing-inspect @@ -597,7 +581,6 @@ voir==0.2.15 # -r benchmarks/flops/requirements.in # -r benchmarks/huggingface/requirements.in # -r benchmarks/llama/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # -r benchmarks/timm/requirements.in diff --git a/.pin/constraints-hpu-torch.txt b/.pin/constraints-hpu-torch.txt index de77f3bfd..4a26a461a 100644 --- a/.pin/constraints-hpu-torch.txt +++ b/.pin/constraints-hpu-torch.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=.pin/constraints-hpu-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in +# pip-compile --output-file=.pin/constraints-hpu-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in # --extra-index-url https://pypi.ngc.nvidia.com --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html @@ -27,6 +27,8 @@ aiohttp==3.9.5 # fsspec aiosignal==1.3.1 # via aiohttp +annotated-types==0.7.0 + # via pydantic antlr4-python3-runtime==4.9.3 # via omegaconf asttokens==2.4.1 @@ -57,9 +59,7 @@ datasets==2.19.2 # -r benchmarks/llama/requirements.in # evaluate deepspeed==0.14.2 - # via - # -r benchmarks/accelerate_opt/requirements.in - # -r benchmarks/rwkv/requirements.in + # via -r benchmarks/accelerate_opt/requirements.in dill==0.3.8 # via # datasets @@ -115,7 +115,6 @@ fsspec[http]==2024.3.1 # etils # evaluate # huggingface-hub - # pytorch-lightning # torch # torchx future==1.0.0 @@ -191,9 +190,7 @@ jinja2==3.1.4 joblib==1.4.2 # via scikit-learn lightning-utilities==0.11.2 - # via - # pytorch-lightning - # torchmetrics + # via torchmetrics markdown==3.6 # via tensorboard markdown-it-py==3.0.0 @@ -238,13 +235,10 @@ nest-asyncio==1.6.0 networkx==3.3 # via torch ninja==1.11.1.1 - # via - # -r benchmarks/rwkv/requirements.in - # deepspeed + # via deepspeed numpy==1.26.4 # via # -r benchmarks/dlrm/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # accelerate @@ -270,7 +264,6 @@ numpy==1.26.4 # orbax-checkpoint # pandas # pyarrow - # pytorch-lightning # scikit-learn # scipy # tensorboard @@ -354,7 +347,6 @@ packaging==24.1 # evaluate # huggingface-hub # lightning-utilities - # pytorch-lightning # tensorboardx # torchmetrics # transformers @@ -385,10 +377,10 @@ pyarrow==16.1.0 # via datasets pyarrow-hotfix==0.6 # via datasets -pydantic==1.10.15 - # via - # -r benchmarks/rwkv/requirements.in - # deepspeed +pydantic==2.7.3 + # via deepspeed +pydantic-core==2.18.4 + # via pydantic pydot==2.0.0 # via -r benchmarks/dlrm/requirements.in pygments==2.18.0 @@ -407,8 +399,6 @@ python-dateutil==2.9.0.post0 # via pandas pytinyrenderer==0.0.14 # via brax -pytorch-lightning==1.9.5 - # via -r benchmarks/rwkv/requirements.in pytz==2024.1 # via pandas pyyaml==6.0.1 @@ -421,7 +411,6 @@ pyyaml==6.0.1 # ml-collections # omegaconf # orbax-checkpoint - # pytorch-lightning # torchx # transformers reactivex==4.0.4 @@ -474,7 +463,7 @@ tensorboard-data-server==0.7.2 # via tensorboard tensorboardx==2.6.2.2 # via brax -tensorstore==0.1.60 +tensorstore==0.1.61 # via # flax # orbax-checkpoint @@ -494,7 +483,6 @@ torch==2.3.1 # -r benchmarks/flops/requirements.in # -r benchmarks/huggingface/requirements.in # -r benchmarks/llama/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # -r benchmarks/timm/requirements.in @@ -503,7 +491,6 @@ torch==2.3.1 # accelerate # deepspeed # fairscale - # pytorch-lightning # torchaudio # torchcompat # torchmetrics @@ -518,9 +505,7 @@ torchcompat==1.0.2 # -r benchmarks/torchvision/requirements.in # -r benchmarks/torchvision_ddp/requirements.in torchmetrics==1.0.3 - # via - # pytorch-lightning - # torchrec + # via torchrec torchrec==0.7.0 # via -r benchmarks/dlrm/requirements.in torchvision==0.18.1 @@ -547,7 +532,6 @@ tqdm==4.66.4 # deepspeed # evaluate # huggingface-hub - # pytorch-lightning # torchrec # transformers transformers==4.41.2 @@ -571,8 +555,8 @@ typing-extensions==4.12.2 # lightning-utilities # orbax-checkpoint # pydantic + # pydantic-core # pyre-extensions - # pytorch-lightning # reactivex # torch # typing-inspect @@ -596,7 +580,6 @@ voir==0.2.15 # -r benchmarks/flops/requirements.in # -r benchmarks/huggingface/requirements.in # -r benchmarks/llama/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # -r benchmarks/timm/requirements.in diff --git a/.pin/constraints-rocm-torch.txt b/.pin/constraints-rocm-torch.txt index 156333c22..472bc62fb 100644 --- a/.pin/constraints-rocm-torch.txt +++ b/.pin/constraints-rocm-torch.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=.pin/constraints-rocm-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in +# pip-compile --output-file=.pin/constraints-rocm-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in # --extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://download.pytorch.org/whl/rocm6.0 @@ -28,6 +28,8 @@ aiohttp==3.9.5 # fsspec aiosignal==1.3.1 # via aiohttp +annotated-types==0.7.0 + # via pydantic antlr4-python3-runtime==4.9.3 # via omegaconf asttokens==2.4.1 @@ -58,9 +60,7 @@ datasets==2.19.2 # -r benchmarks/llama/requirements.in # evaluate deepspeed==0.14.2 - # via - # -r benchmarks/accelerate_opt/requirements.in - # -r benchmarks/rwkv/requirements.in + # via -r benchmarks/accelerate_opt/requirements.in dill==0.3.8 # via # datasets @@ -74,7 +74,7 @@ docker==7.1.0 # via torchx docstring-parser==0.8.1 # via torchx -etils[epath,epy]==1.9.1 +etils[epath,epy]==1.9.2 # via # brax # mujoco @@ -116,7 +116,6 @@ fsspec[http]==2024.3.1 # etils # evaluate # huggingface-hub - # pytorch-lightning # torch # torchx future==1.0.0 @@ -159,7 +158,7 @@ importlib-resources==6.4.0 # torchcompat itsdangerous==2.2.0 # via flask -jax[cuda12]==0.4.28 +jax[cuda12]==0.4.29 # via # -r benchmarks/brax/requirements.in # brax @@ -169,11 +168,11 @@ jax[cuda12]==0.4.28 # mujoco-mjx # optax # orbax-checkpoint -jax-cuda12-pjrt==0.4.28 +jax-cuda12-pjrt==0.4.29 # via jax-cuda12-plugin -jax-cuda12-plugin==0.4.28 +jax-cuda12-plugin==0.4.29 # via jax -jaxlib==0.4.28+cuda12.cudnn89 +jaxlib==0.4.29+cuda12.cudnn91 # via # brax # chex @@ -192,9 +191,7 @@ jinja2==3.1.4 joblib==1.4.2 # via scikit-learn lightning-utilities==0.11.2 - # via - # pytorch-lightning - # torchmetrics + # via torchmetrics markdown==3.6 # via tensorboard markdown-it-py==3.0.0 @@ -239,13 +236,10 @@ nest-asyncio==1.6.0 networkx==3.3 # via torch ninja==1.11.1.1 - # via - # -r benchmarks/rwkv/requirements.in - # deepspeed + # via deepspeed numpy==1.26.4 # via # -r benchmarks/dlrm/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # accelerate @@ -271,7 +265,6 @@ numpy==1.26.4 # orbax-checkpoint # pandas # pyarrow - # pytorch-lightning # scikit-learn # scipy # tensorboard @@ -290,11 +283,9 @@ nvidia-cuda-cupti-cu12==12.5.39 # via jax nvidia-cuda-nvcc-cu12==12.5.40 # via jax -nvidia-cuda-nvrtc-cu12==12.5.40 - # via nvidia-cudnn-cu12 nvidia-cuda-runtime-cu12==12.5.39 # via jax -nvidia-cudnn-cu12==8.9.7.29 +nvidia-cudnn-cu12==9.1.1.17 # via jax nvidia-cufft-cu12==11.2.3.18 # via jax @@ -324,7 +315,7 @@ optax==0.2.2 # via # brax # flax -orbax-checkpoint==0.5.15 +orbax-checkpoint==0.5.16 # via # brax # flax @@ -338,7 +329,6 @@ packaging==24.1 # evaluate # huggingface-hub # lightning-utilities - # pytorch-lightning # tensorboardx # torchmetrics # transformers @@ -369,10 +359,10 @@ pyarrow==16.1.0 # via datasets pyarrow-hotfix==0.6 # via datasets -pydantic==1.10.15 - # via - # -r benchmarks/rwkv/requirements.in - # deepspeed +pydantic==2.7.3 + # via deepspeed +pydantic-core==2.18.4 + # via pydantic pydot==2.0.0 # via -r benchmarks/dlrm/requirements.in pygments==2.18.0 @@ -391,8 +381,6 @@ python-dateutil==2.9.0.post0 # via pandas pytinyrenderer==0.0.14 # via brax -pytorch-lightning==1.9.5 - # via -r benchmarks/rwkv/requirements.in pytorch-triton-rocm==2.3.1 # via torch pytz==2024.1 @@ -407,7 +395,6 @@ pyyaml==6.0.1 # ml-collections # omegaconf # orbax-checkpoint - # pytorch-lightning # torchx # transformers reactivex==4.0.4 @@ -460,7 +447,7 @@ tensorboard-data-server==0.7.2 # via tensorboard tensorboardx==2.6.2.2 # via brax -tensorstore==0.1.60 +tensorstore==0.1.61 # via # flax # orbax-checkpoint @@ -480,7 +467,6 @@ torch==2.3.1+rocm6.0 # -r benchmarks/flops/requirements.in # -r benchmarks/huggingface/requirements.in # -r benchmarks/llama/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # -r benchmarks/timm/requirements.in @@ -489,7 +475,6 @@ torch==2.3.1+rocm6.0 # accelerate # deepspeed # fairscale - # pytorch-lightning # torchaudio # torchcompat # torchmetrics @@ -504,9 +489,7 @@ torchcompat==1.0.2 # -r benchmarks/torchvision/requirements.in # -r benchmarks/torchvision_ddp/requirements.in torchmetrics==1.0.3 - # via - # pytorch-lightning - # torchrec + # via torchrec torchrec==0.7.0 # via -r benchmarks/dlrm/requirements.in torchvision==0.18.1+rocm6.0 @@ -533,7 +516,6 @@ tqdm==4.66.4 # deepspeed # evaluate # huggingface-hub - # pytorch-lightning # torchrec # transformers transformers==4.41.2 @@ -555,8 +537,8 @@ typing-extensions==4.12.2 # lightning-utilities # orbax-checkpoint # pydantic + # pydantic-core # pyre-extensions - # pytorch-lightning # reactivex # torch # typing-inspect @@ -580,7 +562,6 @@ voir==0.2.15 # -r benchmarks/flops/requirements.in # -r benchmarks/huggingface/requirements.in # -r benchmarks/llama/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # -r benchmarks/timm/requirements.in diff --git a/.pin/constraints-xpu-torch.txt b/.pin/constraints-xpu-torch.txt index 1ffcf44a1..f33653b16 100644 --- a/.pin/constraints-xpu-torch.txt +++ b/.pin/constraints-xpu-torch.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=.pin/constraints-xpu-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in +# pip-compile --output-file=.pin/constraints-xpu-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in # --extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ @@ -28,6 +28,8 @@ aiohttp==3.9.5 # fsspec aiosignal==1.3.1 # via aiohttp +annotated-types==0.7.0 + # via pydantic antlr4-python3-runtime==4.9.3 # via omegaconf asttokens==2.4.1 @@ -52,16 +54,14 @@ codefind==0.1.6 # via ptera contextlib2==21.6.0 # via ml-collections -datasets==2.19.2 +datasets==2.14.7 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/llama/requirements.in # evaluate deepspeed==0.14.2 - # via - # -r benchmarks/accelerate_opt/requirements.in - # -r benchmarks/rwkv/requirements.in -dill==0.3.8 + # via -r benchmarks/accelerate_opt/requirements.in +dill==0.3.7 # via # datasets # evaluate @@ -72,7 +72,7 @@ dm-tree==0.1.8 # via dm-env docker==7.1.0 # via torchx -docstring-parser==0.8.1 +docstring-parser==0.16 # via torchx etils[epath,epy]==1.9.1 # via @@ -90,7 +90,6 @@ fbgemm-gpu==0.7.0 # via torchrec filelock==3.14.0 # via - # datasets # huggingface-hub # torch # torchx @@ -110,13 +109,12 @@ frozenlist==1.4.1 # via # aiohttp # aiosignal -fsspec[http]==2024.3.1 +fsspec[http]==2023.1.0 # via # datasets # etils # evaluate # huggingface-hub - # pytorch-lightning # torch # torchx future==1.0.0 @@ -139,7 +137,7 @@ gym-notices==0.0.8 # via gym hjson==3.1.0 # via deepspeed -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -r benchmarks/timm/requirements.in # accelerate @@ -192,9 +190,7 @@ jinja2==3.1.4 joblib==1.4.2 # via scikit-learn lightning-utilities==0.11.2 - # via - # pytorch-lightning - # torchmetrics + # via torchmetrics markdown==3.6 # via tensorboard markdown-it-py==3.0.0 @@ -228,7 +224,7 @@ multidict==6.0.5 # via # aiohttp # yarl -multiprocess==0.70.16 +multiprocess==0.70.15 # via # datasets # evaluate @@ -239,13 +235,10 @@ nest-asyncio==1.6.0 networkx==3.3 # via torch ninja==1.11.1.1 - # via - # -r benchmarks/rwkv/requirements.in - # deepspeed + # via deepspeed numpy==1.26.4 # via # -r benchmarks/dlrm/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # accelerate @@ -271,7 +264,6 @@ numpy==1.26.4 # orbax-checkpoint # pandas # pyarrow - # pytorch-lightning # scikit-learn # scipy # tensorboard @@ -355,7 +347,6 @@ packaging==24.1 # evaluate # huggingface-hub # lightning-utilities - # pytorch-lightning # tensorboardx # torchmetrics # transformers @@ -386,10 +377,10 @@ pyarrow==16.1.0 # via datasets pyarrow-hotfix==0.6 # via datasets -pydantic==1.10.15 - # via - # -r benchmarks/rwkv/requirements.in - # deepspeed +pydantic==2.7.3 + # via deepspeed +pydantic-core==2.18.4 + # via pydantic pydot==2.0.0 # via -r benchmarks/dlrm/requirements.in pygments==2.18.0 @@ -408,8 +399,6 @@ python-dateutil==2.9.0.post0 # via pandas pytinyrenderer==0.0.14 # via brax -pytorch-lightning==1.9.5 - # via -r benchmarks/rwkv/requirements.in pytz==2024.1 # via pandas pyyaml==6.0.1 @@ -422,7 +411,6 @@ pyyaml==6.0.1 # ml-collections # omegaconf # orbax-checkpoint - # pytorch-lightning # torchx # transformers reactivex==4.0.4 @@ -434,6 +422,7 @@ requests==2.32.3 # datasets # docker # evaluate + # fsspec # huggingface-hub # transformers rich==13.7.1 @@ -475,7 +464,7 @@ tensorboard-data-server==0.7.2 # via tensorboard tensorboardx==2.6.2.2 # via brax -tensorstore==0.1.60 +tensorstore==0.1.61 # via # flax # orbax-checkpoint @@ -483,7 +472,7 @@ termcolor==2.4.0 # via fire threadpoolctl==3.5.0 # via scikit-learn -tokenizers==0.19.1 +tokenizers==0.15.2 # via transformers toolz==0.12.1 # via chex @@ -496,7 +485,6 @@ torch==2.3.1 # -r benchmarks/flops/requirements.in # -r benchmarks/huggingface/requirements.in # -r benchmarks/llama/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # -r benchmarks/timm/requirements.in @@ -505,7 +493,6 @@ torch==2.3.1 # accelerate # deepspeed # fairscale - # pytorch-lightning # torchaudio # torchcompat # torchmetrics @@ -522,9 +509,7 @@ torchcompat==1.0.2 # -r benchmarks/torchvision/requirements.in # -r benchmarks/torchvision_ddp/requirements.in torchmetrics==1.0.3 - # via - # pytorch-lightning - # torchrec + # via torchrec torchrec==0.7.0 # via -r benchmarks/dlrm/requirements.in torchvision==0.18.1 @@ -539,7 +524,7 @@ torchvision==0.18.1 # -r benchmarks/torchvision_ddp/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in -torchx==0.5.0 +torchx==0.6.0 # via -r benchmarks/dlrm/requirements.in tqdm==4.66.4 # via @@ -552,10 +537,9 @@ tqdm==4.66.4 # deepspeed # evaluate # huggingface-hub - # pytorch-lightning # torchrec # transformers -transformers==4.41.2 +transformers==4.35.2 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/huggingface/requirements.in @@ -576,8 +560,8 @@ typing-extensions==4.12.2 # lightning-utilities # orbax-checkpoint # pydantic + # pydantic-core # pyre-extensions - # pytorch-lightning # reactivex # torch # typing-inspect @@ -601,7 +585,6 @@ voir==0.2.15 # -r benchmarks/flops/requirements.in # -r benchmarks/huggingface/requirements.in # -r benchmarks/llama/requirements.in - # -r benchmarks/rwkv/requirements.in # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # -r benchmarks/timm/requirements.in diff --git a/benchmarks/accelerate_opt/requirements.cuda.txt b/benchmarks/accelerate_opt/requirements.cuda.txt index 76e10219d..5848f226c 100644 --- a/benchmarks/accelerate_opt/requirements.cuda.txt +++ b/benchmarks/accelerate_opt/requirements.cuda.txt @@ -22,6 +22,10 @@ aiosignal==1.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp +annotated-types==0.7.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pydantic antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -46,7 +50,7 @@ codefind==0.1.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera -datasets==2.19.2 +datasets==2.14.7 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/accelerate_opt/requirements.in @@ -55,7 +59,7 @@ deepspeed==0.14.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/accelerate_opt/requirements.in -dill==0.3.8 +dill==0.3.7 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -72,7 +76,6 @@ executing==1.2.0 filelock==3.14.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # datasets # huggingface-hub # torch # transformers @@ -82,7 +85,7 @@ frozenlist==1.4.1 # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # aiosignal -fsspec[http]==2024.3.1 +fsspec[http]==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -98,7 +101,7 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # accelerate @@ -136,7 +139,7 @@ multidict==6.0.5 # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # yarl -multiprocess==0.70.16 +multiprocess==0.70.15 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -260,10 +263,14 @@ pyarrow-hotfix==0.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets -pydantic==1.10.15 +pydantic==2.7.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed +pydantic-core==2.18.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pydantic pygments==2.18.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -302,6 +309,7 @@ requests==2.32.3 # -c .pin/../.pin/constraints-cuda-torch.txt # datasets # evaluate + # fsspec # huggingface-hub # transformers rich==13.7.1 @@ -323,7 +331,7 @@ sympy==1.12.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -tokenizers==0.19.1 +tokenizers==0.15.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # transformers @@ -351,7 +359,7 @@ tqdm==4.66.4 # evaluate # huggingface-hub # transformers -transformers==4.41.2 +transformers==4.35.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/accelerate_opt/requirements.in @@ -364,6 +372,7 @@ typing-extensions==4.12.2 # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # pydantic + # pydantic-core # reactivex # torch tzdata==2024.1 @@ -379,7 +388,10 @@ varname==0.10.0 # -c .pin/../.pin/constraints-cuda-torch.txt # giving voir==0.2.15 - # via -r benchmarks/accelerate_opt/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/accelerate_opt/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt diff --git a/benchmarks/accelerate_opt/requirements.hpu.txt b/benchmarks/accelerate_opt/requirements.hpu.txt index fc464bcaf..ed65d2ced 100644 --- a/benchmarks/accelerate_opt/requirements.hpu.txt +++ b/benchmarks/accelerate_opt/requirements.hpu.txt @@ -21,6 +21,10 @@ aiosignal==1.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # aiohttp +annotated-types==0.7.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # pydantic antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -259,10 +263,14 @@ pyarrow-hotfix==0.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets -pydantic==1.10.15 +pydantic==2.7.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed +pydantic-core==2.18.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # pydantic pygments==2.18.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -363,6 +371,7 @@ typing-extensions==4.12.2 # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub # pydantic + # pydantic-core # reactivex # torch tzdata==2024.1 @@ -378,7 +387,10 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/accelerate_opt/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/accelerate_opt/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index eb5902c86..ca23d9979 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -22,6 +22,10 @@ aiosignal==1.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp +annotated-types==0.7.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pydantic antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -208,10 +212,14 @@ pyarrow-hotfix==0.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets -pydantic==1.10.15 +pydantic==2.7.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed +pydantic-core==2.18.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pydantic pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -312,6 +320,7 @@ typing-extensions==4.12.2 # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pydantic + # pydantic-core # reactivex # torch tzdata==2024.1 @@ -327,7 +336,10 @@ varname==0.10.0 # -c .pin/../.pin/constraints-rocm-torch.txt # giving voir==0.2.15 - # via -r benchmarks/accelerate_opt/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/accelerate_opt/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt diff --git a/benchmarks/accelerate_opt/requirements.xpu.txt b/benchmarks/accelerate_opt/requirements.xpu.txt index 0d34ed0f0..897df1173 100644 --- a/benchmarks/accelerate_opt/requirements.xpu.txt +++ b/benchmarks/accelerate_opt/requirements.xpu.txt @@ -22,6 +22,10 @@ aiosignal==1.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp +annotated-types==0.7.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pydantic antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -46,7 +50,7 @@ codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera -datasets==2.19.2 +datasets==2.14.7 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/accelerate_opt/requirements.in @@ -55,7 +59,7 @@ deepspeed==0.14.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/accelerate_opt/requirements.in -dill==0.3.8 +dill==0.3.7 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -72,7 +76,6 @@ executing==1.2.0 filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # datasets # huggingface-hub # torch # transformers @@ -82,7 +85,7 @@ frozenlist==1.4.1 # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp # aiosignal -fsspec[http]==2024.3.1 +fsspec[http]==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -98,7 +101,7 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # deepspeed -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # accelerate @@ -136,7 +139,7 @@ multidict==6.0.5 # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp # yarl -multiprocess==0.70.16 +multiprocess==0.70.15 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -260,10 +263,14 @@ pyarrow-hotfix==0.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets -pydantic==1.10.15 +pydantic==2.7.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # deepspeed +pydantic-core==2.18.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pydantic pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -302,6 +309,7 @@ requests==2.32.3 # -c .pin/../.pin/constraints-xpu-torch.txt # datasets # evaluate + # fsspec # huggingface-hub # transformers rich==13.7.1 @@ -323,7 +331,7 @@ sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -tokenizers==0.19.1 +tokenizers==0.15.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # transformers @@ -354,7 +362,7 @@ tqdm==4.66.4 # evaluate # huggingface-hub # transformers -transformers==4.41.2 +transformers==4.35.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/accelerate_opt/requirements.in @@ -367,6 +375,7 @@ typing-extensions==4.12.2 # -c .pin/../.pin/constraints-xpu-torch.txt # huggingface-hub # pydantic + # pydantic-core # reactivex # torch tzdata==2024.1 diff --git a/benchmarks/brax/requirements.cuda.txt b/benchmarks/brax/requirements.cuda.txt index 38a54509a..521176f1f 100644 --- a/benchmarks/brax/requirements.cuda.txt +++ b/benchmarks/brax/requirements.cuda.txt @@ -93,7 +93,7 @@ flax==0.8.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # brax -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # etils @@ -396,7 +396,7 @@ tensorboardx==2.6.2.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # brax -tensorstore==0.1.60 +tensorstore==0.1.61 # via # -c .pin/../.pin/constraints-cuda-torch.txt # flax diff --git a/benchmarks/brax/requirements.hpu.txt b/benchmarks/brax/requirements.hpu.txt index d79e7242c..b2fcd3de6 100644 --- a/benchmarks/brax/requirements.hpu.txt +++ b/benchmarks/brax/requirements.hpu.txt @@ -395,7 +395,7 @@ tensorboardx==2.6.2.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # brax -tensorstore==0.1.60 +tensorstore==0.1.61 # via # -c .pin/../.pin/constraints-hpu-torch.txt # flax diff --git a/benchmarks/brax/requirements.rocm.txt b/benchmarks/brax/requirements.rocm.txt index c77018b5e..8c3b6f67b 100644 --- a/benchmarks/brax/requirements.rocm.txt +++ b/benchmarks/brax/requirements.rocm.txt @@ -64,7 +64,7 @@ dm-tree==0.1.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # dm-env -etils[epath,epy]==1.9.1 +etils[epath,epy]==1.9.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax @@ -127,7 +127,7 @@ itsdangerous==2.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -jax[cuda12]==0.4.28 +jax[cuda12]==0.4.29 # via # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/brax/requirements.in @@ -138,15 +138,15 @@ jax[cuda12]==0.4.28 # mujoco-mjx # optax # orbax-checkpoint -jax-cuda12-pjrt==0.4.28 +jax-cuda12-pjrt==0.4.29 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax-cuda12-plugin -jax-cuda12-plugin==0.4.28 +jax-cuda12-plugin==0.4.29 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax -jaxlib==0.4.28+cuda12.cudnn89 +jaxlib==0.4.29+cuda12.cudnn91 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax @@ -249,15 +249,11 @@ nvidia-cuda-nvcc-cu12==12.5.40 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax -nvidia-cuda-nvrtc-cu12==12.5.40 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # nvidia-cudnn-cu12 nvidia-cuda-runtime-cu12==12.5.39 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax -nvidia-cudnn-cu12==8.9.7.29 +nvidia-cudnn-cu12==9.1.1.17 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax @@ -298,7 +294,7 @@ optax==0.2.2 # -c .pin/../.pin/constraints-rocm-torch.txt # brax # flax -orbax-checkpoint==0.5.15 +orbax-checkpoint==0.5.16 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax @@ -385,7 +381,7 @@ tensorboardx==2.6.2.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -tensorstore==0.1.60 +tensorstore==0.1.61 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax diff --git a/benchmarks/brax/requirements.xpu.txt b/benchmarks/brax/requirements.xpu.txt index 75a03f5aa..7ce6e5df8 100644 --- a/benchmarks/brax/requirements.xpu.txt +++ b/benchmarks/brax/requirements.xpu.txt @@ -93,7 +93,7 @@ flax==0.8.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # brax -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # etils @@ -396,7 +396,7 @@ tensorboardx==2.6.2.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # brax -tensorstore==0.1.60 +tensorstore==0.1.61 # via # -c .pin/../.pin/constraints-xpu-torch.txt # flax diff --git a/benchmarks/dlrm/requirements.cuda.txt b/benchmarks/dlrm/requirements.cuda.txt index b8e79dd60..c97f4fac8 100644 --- a/benchmarks/dlrm/requirements.cuda.txt +++ b/benchmarks/dlrm/requirements.cuda.txt @@ -37,7 +37,7 @@ docker==7.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchx -docstring-parser==0.8.1 +docstring-parser==0.16 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchx @@ -55,7 +55,7 @@ filelock==3.14.0 # torch # torchx # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch @@ -306,7 +306,7 @@ torchviz==0.0.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/dlrm/requirements.in -torchx==0.5.0 +torchx==0.6.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/dlrm/requirements.in diff --git a/benchmarks/dlrm/requirements.hpu.txt b/benchmarks/dlrm/requirements.hpu.txt index 1c7902cf6..fb562c128 100644 --- a/benchmarks/dlrm/requirements.hpu.txt +++ b/benchmarks/dlrm/requirements.hpu.txt @@ -341,7 +341,10 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/dlrm/requirements.in werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/dlrm/requirements.xpu.txt b/benchmarks/dlrm/requirements.xpu.txt index e26b4bbac..c6b38d2ea 100644 --- a/benchmarks/dlrm/requirements.xpu.txt +++ b/benchmarks/dlrm/requirements.xpu.txt @@ -37,7 +37,7 @@ docker==7.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchx -docstring-parser==0.8.1 +docstring-parser==0.16 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchx @@ -55,7 +55,7 @@ filelock==3.14.0 # torch # torchx # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -307,7 +307,7 @@ torchviz==0.0.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in -torchx==0.5.0 +torchx==0.6.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in diff --git a/benchmarks/flops/requirements.cuda.txt b/benchmarks/flops/requirements.cuda.txt index 21a41a149..f243aa00b 100644 --- a/benchmarks/flops/requirements.cuda.txt +++ b/benchmarks/flops/requirements.cuda.txt @@ -30,7 +30,7 @@ filelock==3.14.0 # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch @@ -204,4 +204,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-cuda-torch.txt # giving voir==0.2.15 - # via -r benchmarks/flops/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.hpu.txt b/benchmarks/flops/requirements.hpu.txt index 08eda7ab3..b7b7eff16 100644 --- a/benchmarks/flops/requirements.hpu.txt +++ b/benchmarks/flops/requirements.hpu.txt @@ -203,4 +203,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/flops/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.rocm.txt b/benchmarks/flops/requirements.rocm.txt index ff65456a0..e7a099d1b 100644 --- a/benchmarks/flops/requirements.rocm.txt +++ b/benchmarks/flops/requirements.rocm.txt @@ -152,4 +152,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-rocm-torch.txt # giving voir==0.2.15 - # via -r benchmarks/flops/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.xpu.txt b/benchmarks/flops/requirements.xpu.txt index 0fc5be0db..2ae9bf4eb 100644 --- a/benchmarks/flops/requirements.xpu.txt +++ b/benchmarks/flops/requirements.xpu.txt @@ -30,7 +30,7 @@ filelock==3.14.0 # -c .pin/../.pin/constraints-xpu-torch.txt # torch # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch diff --git a/benchmarks/huggingface/requirements.cuda.txt b/benchmarks/huggingface/requirements.cuda.txt index e095395ee..a8b916274 100644 --- a/benchmarks/huggingface/requirements.cuda.txt +++ b/benchmarks/huggingface/requirements.cuda.txt @@ -40,7 +40,7 @@ filelock==3.14.0 # torch # transformers # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub @@ -50,7 +50,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # tokenizers @@ -203,7 +203,7 @@ sympy==1.12.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -tokenizers==0.19.1 +tokenizers==0.15.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # transformers @@ -216,7 +216,7 @@ tqdm==4.66.4 # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # transformers -transformers==4.41.2 +transformers==4.35.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/huggingface/requirements.in @@ -239,4 +239,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-cuda-torch.txt # giving voir==0.2.15 - # via -r benchmarks/huggingface/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.hpu.txt b/benchmarks/huggingface/requirements.hpu.txt index e0190abae..bd7206fd9 100644 --- a/benchmarks/huggingface/requirements.hpu.txt +++ b/benchmarks/huggingface/requirements.hpu.txt @@ -238,4 +238,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/huggingface/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.rocm.txt b/benchmarks/huggingface/requirements.rocm.txt index a84b09b52..b22d9dd3e 100644 --- a/benchmarks/huggingface/requirements.rocm.txt +++ b/benchmarks/huggingface/requirements.rocm.txt @@ -187,4 +187,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-rocm-torch.txt # giving voir==0.2.15 - # via -r benchmarks/huggingface/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.xpu.txt b/benchmarks/huggingface/requirements.xpu.txt index f46a61603..e85bf3e9c 100644 --- a/benchmarks/huggingface/requirements.xpu.txt +++ b/benchmarks/huggingface/requirements.xpu.txt @@ -40,7 +40,7 @@ filelock==3.14.0 # torch # transformers # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # huggingface-hub @@ -50,7 +50,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # tokenizers @@ -203,7 +203,7 @@ sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -tokenizers==0.19.1 +tokenizers==0.15.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # transformers @@ -217,7 +217,7 @@ tqdm==4.66.4 # -c .pin/../.pin/constraints-xpu-torch.txt # huggingface-hub # transformers -transformers==4.41.2 +transformers==4.35.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/llama/main.py b/benchmarks/llama/main.py index da00a9c91..499bd7263 100755 --- a/benchmarks/llama/main.py +++ b/benchmarks/llama/main.py @@ -178,7 +178,7 @@ def main(): if __name__ == "__main__": - from voir.wrapper import StopProgram + from benchmate.metrics import StopProgram import traceback try: main() diff --git a/benchmarks/llama/requirements.cuda.txt b/benchmarks/llama/requirements.cuda.txt index 52a3b8c66..1d156ed34 100644 --- a/benchmarks/llama/requirements.cuda.txt +++ b/benchmarks/llama/requirements.cuda.txt @@ -42,11 +42,11 @@ codefind==0.1.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera -datasets==2.19.2 +datasets==2.14.7 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/llama/requirements.in -dill==0.3.8 +dill==0.3.7 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -62,7 +62,6 @@ fairscale==0.4.13 filelock==3.14.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # datasets # huggingface-hub # torch # transformers @@ -76,7 +75,7 @@ frozenlist==1.4.1 # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # aiosignal -fsspec[http]==2024.3.1 +fsspec[http]==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -87,7 +86,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -123,7 +122,7 @@ multidict==6.0.5 # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # yarl -multiprocess==0.70.16 +multiprocess==0.70.15 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -260,6 +259,7 @@ requests==2.32.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets + # fsspec # huggingface-hub # transformers rich==13.7.1 @@ -288,7 +288,7 @@ termcolor==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # fire -tokenizers==0.19.1 +tokenizers==0.15.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # transformers @@ -303,7 +303,7 @@ tqdm==4.66.4 # datasets # huggingface-hub # transformers -transformers==4.41.2 +transformers==4.35.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/llama/requirements.in @@ -326,9 +326,14 @@ urllib3==1.26.18 # -c .pin/../.pin/constraints-cuda-torch.txt # requests varname==0.10.0 - # via giving + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving voir==0.2.15 - # via -r benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/llama/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt diff --git a/benchmarks/llama/requirements.hpu.txt b/benchmarks/llama/requirements.hpu.txt index d4b7b2514..ff0f7f60b 100644 --- a/benchmarks/llama/requirements.hpu.txt +++ b/benchmarks/llama/requirements.hpu.txt @@ -329,7 +329,10 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/llama/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/llama/requirements.rocm.txt b/benchmarks/llama/requirements.rocm.txt index d7d177dba..bcd7599ca 100644 --- a/benchmarks/llama/requirements.rocm.txt +++ b/benchmarks/llama/requirements.rocm.txt @@ -274,9 +274,14 @@ urllib3==1.26.18 # -c .pin/../.pin/constraints-rocm-torch.txt # requests varname==0.10.0 - # via giving + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving voir==0.2.15 - # via -r benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/llama/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt diff --git a/benchmarks/llama/requirements.xpu.txt b/benchmarks/llama/requirements.xpu.txt index cb7355a26..789c44ff4 100644 --- a/benchmarks/llama/requirements.xpu.txt +++ b/benchmarks/llama/requirements.xpu.txt @@ -42,11 +42,11 @@ codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera -datasets==2.19.2 +datasets==2.14.7 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/llama/requirements.in -dill==0.3.8 +dill==0.3.7 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -62,7 +62,6 @@ fairscale==0.4.13 filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # datasets # huggingface-hub # torch # transformers @@ -76,7 +75,7 @@ frozenlist==1.4.1 # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp # aiosignal -fsspec[http]==2024.3.1 +fsspec[http]==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -87,7 +86,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -123,7 +122,7 @@ multidict==6.0.5 # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp # yarl -multiprocess==0.70.16 +multiprocess==0.70.15 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -260,6 +259,7 @@ requests==2.32.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets + # fsspec # huggingface-hub # transformers rich==13.7.1 @@ -288,7 +288,7 @@ termcolor==2.4.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # fire -tokenizers==0.19.1 +tokenizers==0.15.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # transformers @@ -304,7 +304,7 @@ tqdm==4.66.4 # datasets # huggingface-hub # transformers -transformers==4.41.2 +transformers==4.35.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/llama/requirements.in diff --git a/benchmarks/stargan/prepare.py b/benchmarks/stargan/prepare.py index c17e38033..dcb54a0f5 100755 --- a/benchmarks/stargan/prepare.py +++ b/benchmarks/stargan/prepare.py @@ -1,7 +1,11 @@ +#!/usr/bin/env python +import os - -def download_celebA(): +def download_huggingface_celebA(): + # + # Format is no good + # from datasets import load_dataset dataset = load_dataset( "student/celebA", @@ -10,5 +14,30 @@ def download_celebA(): ) +def download_torchvision_celebA(): + # + # pip install gdown + # + # gdown.exceptions.FileURLRetrievalError: Failed to retrieve file url: + # + # Too many users have viewed or downloaded this file recently. Please + # try accessing the file again later. If the file you are trying to + # access is particularly large or is shared with many people, it may + # take up to 24 hours to be able to view or download the file. If you + # still can't access a file after 24 hours, contact your domain + # administrator. + + from torchvision.datasets import CelebA + dataset = CelebA( + os.path.join(os.environ["MILABENCH_DIR_DATA"], "CelebA"), + split="train", + download=True + ) + +def main(): + """""" + # download_torchvision_celebA() + + if __name__ == "__main__": - download_celebA() + main() diff --git a/benchmarks/stargan/requirements.cuda.txt b/benchmarks/stargan/requirements.cuda.txt index d2541bf52..46b58873e 100644 --- a/benchmarks/stargan/requirements.cuda.txt +++ b/benchmarks/stargan/requirements.cuda.txt @@ -30,7 +30,7 @@ filelock==3.14.0 # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch @@ -191,4 +191,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-cuda-torch.txt # giving voir==0.2.15 - # via -r benchmarks/stargan/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/stargan/requirements.in diff --git a/benchmarks/stargan/requirements.hpu.txt b/benchmarks/stargan/requirements.hpu.txt index 10ad3e3eb..da836f25f 100644 --- a/benchmarks/stargan/requirements.hpu.txt +++ b/benchmarks/stargan/requirements.hpu.txt @@ -190,4 +190,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/stargan/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/stargan/requirements.in diff --git a/benchmarks/stargan/requirements.in b/benchmarks/stargan/requirements.in index bae650375..0656da948 100644 --- a/benchmarks/stargan/requirements.in +++ b/benchmarks/stargan/requirements.in @@ -2,3 +2,4 @@ numpy torch torchvision voir +gdown diff --git a/benchmarks/stargan/requirements.rocm.txt b/benchmarks/stargan/requirements.rocm.txt index 0ee10cc6f..a989b8e9d 100644 --- a/benchmarks/stargan/requirements.rocm.txt +++ b/benchmarks/stargan/requirements.rocm.txt @@ -139,4 +139,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-rocm-torch.txt # giving voir==0.2.15 - # via -r benchmarks/stargan/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/stargan/requirements.in diff --git a/benchmarks/stargan/requirements.xpu.txt b/benchmarks/stargan/requirements.xpu.txt index 0ec9b9954..57e5af9f0 100644 --- a/benchmarks/stargan/requirements.xpu.txt +++ b/benchmarks/stargan/requirements.xpu.txt @@ -30,7 +30,7 @@ filelock==3.14.0 # -c .pin/../.pin/constraints-xpu-torch.txt # torch # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch diff --git a/benchmarks/super-slomo/requirements.cuda.txt b/benchmarks/super-slomo/requirements.cuda.txt index df60c83e3..f5b6b26fc 100644 --- a/benchmarks/super-slomo/requirements.cuda.txt +++ b/benchmarks/super-slomo/requirements.cuda.txt @@ -30,7 +30,7 @@ filelock==3.14.0 # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch @@ -200,4 +200,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-cuda-torch.txt # giving voir==0.2.15 - # via -r benchmarks/super-slomo/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.hpu.txt b/benchmarks/super-slomo/requirements.hpu.txt index 993c5aa57..84eb02906 100644 --- a/benchmarks/super-slomo/requirements.hpu.txt +++ b/benchmarks/super-slomo/requirements.hpu.txt @@ -199,4 +199,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/super-slomo/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.rocm.txt b/benchmarks/super-slomo/requirements.rocm.txt index 507bc7012..0914b2e76 100644 --- a/benchmarks/super-slomo/requirements.rocm.txt +++ b/benchmarks/super-slomo/requirements.rocm.txt @@ -148,4 +148,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-rocm-torch.txt # giving voir==0.2.15 - # via -r benchmarks/super-slomo/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.xpu.txt b/benchmarks/super-slomo/requirements.xpu.txt index fccf5db12..a4c5768dc 100644 --- a/benchmarks/super-slomo/requirements.xpu.txt +++ b/benchmarks/super-slomo/requirements.xpu.txt @@ -30,7 +30,7 @@ filelock==3.14.0 # -c .pin/../.pin/constraints-xpu-torch.txt # torch # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch diff --git a/benchmarks/timm/requirements.cuda.txt b/benchmarks/timm/requirements.cuda.txt index 8b162dd50..bcb193e27 100644 --- a/benchmarks/timm/requirements.cuda.txt +++ b/benchmarks/timm/requirements.cuda.txt @@ -39,7 +39,7 @@ filelock==3.14.0 # huggingface-hub # torch # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub @@ -49,7 +49,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/timm/requirements.in @@ -231,4 +231,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-cuda-torch.txt # giving voir==0.2.15 - # via -r benchmarks/timm/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.hpu.txt b/benchmarks/timm/requirements.hpu.txt index ce479d3e4..3d1abfe22 100644 --- a/benchmarks/timm/requirements.hpu.txt +++ b/benchmarks/timm/requirements.hpu.txt @@ -230,4 +230,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/timm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.rocm.txt b/benchmarks/timm/requirements.rocm.txt index 762371c35..5443c2d11 100644 --- a/benchmarks/timm/requirements.rocm.txt +++ b/benchmarks/timm/requirements.rocm.txt @@ -179,4 +179,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-rocm-torch.txt # giving voir==0.2.15 - # via -r benchmarks/timm/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.xpu.txt b/benchmarks/timm/requirements.xpu.txt index 1e32e7820..523a95514 100644 --- a/benchmarks/timm/requirements.xpu.txt +++ b/benchmarks/timm/requirements.xpu.txt @@ -39,7 +39,7 @@ filelock==3.14.0 # huggingface-hub # torch # triton -fsspec==2024.3.1 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # huggingface-hub @@ -49,7 +49,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -huggingface-hub==0.23.3 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/timm/requirements.in diff --git a/benchmarks/torchvision/requirements.cuda.txt b/benchmarks/torchvision/requirements.cuda.txt index 385caea2c..7ebda495d 100644 --- a/benchmarks/torchvision/requirements.cuda.txt +++ b/benchmarks/torchvision/requirements.cuda.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision/requirements.cuda.txt .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision/requirements.in # ---extra-index-url https://download.pytorch.org/whl/cu118 +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera @@ -30,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -fsspec==2023.10.0 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch @@ -44,11 +39,11 @@ giving==0.4.2 # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch @@ -56,7 +51,7 @@ markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -68,31 +63,87 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -108,11 +159,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir @@ -120,34 +167,44 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -torch==2.1.0+cu118 +torch==2.3.1+cu121 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # -r benchmarks/torchvision/requirements.in + # torchcompat # torchvision -torchvision==0.16.0+cu118 - # via -r benchmarks/torchvision/requirements.in -tqdm==4.66.1 - # via -r benchmarks/torchvision/requirements.in -triton==2.1.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # torch -typing-extensions==4.8.0 + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving voir==0.2.15 - # via -r benchmarks/torchvision/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.hpu.txt b/benchmarks/torchvision/requirements.hpu.txt index 21560a4bd..ae2ce864e 100644 --- a/benchmarks/torchvision/requirements.hpu.txt +++ b/benchmarks/torchvision/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/torchvision/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-torchvision.txt milabench/benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision/requirements.hpu.txt .pin/tmp-constraints-hpu-torchvision.txt benchmarks/torchvision/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -109,7 +113,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -162,25 +166,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/torchvision/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in + # torchcompat # torchvision -torchcompat==0.0.1 - # via -r milabench/benchmarks/torchvision/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/torchvision/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/torchvision/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -190,4 +203,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/torchvision/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.rocm.txt b/benchmarks/torchvision/requirements.rocm.txt index f9bb1d4c5..53b4f2d99 100644 --- a/benchmarks/torchvision/requirements.rocm.txt +++ b/benchmarks/torchvision/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision/requirements.rocm.txt .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,19 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,11 +63,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,19 +75,23 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +99,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +111,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,31 +119,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/torchvision/requirements.in - # pytorch-triton-rocm + # torchcompat # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/torchvision/requirements.in -tqdm==4.66.1 - # via -r benchmarks/torchvision/requirements.in -typing-extensions==4.8.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # -r benchmarks/torchvision/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving voir==0.2.15 - # via -r benchmarks/torchvision/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.xpu.txt b/benchmarks/torchvision/requirements.xpu.txt index 0e529ce1f..4546f629f 100644 --- a/benchmarks/torchvision/requirements.xpu.txt +++ b/benchmarks/torchvision/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/torchvision/requirements.xpu.txt .pin/tmp-constraints-xpu-torchvision.txt benchmarks/torchvision/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera @@ -30,11 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -fsspec==2024.2.0 + # triton +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -43,11 +39,11 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -idna==3.7 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -jinja2==3.1.3 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -67,7 +63,7 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # sympy -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -75,6 +71,58 @@ numpy==1.26.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -87,11 +135,15 @@ pillow==10.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -107,10 +159,6 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # torchvision rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -119,34 +167,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in + # torchcompat # torchvision -torchvision==0.16.0a0+cxx11.abi +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in -tqdm==4.66.2 +torchvision==0.18.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in -typing-extensions==4.11.0 +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/torchvision/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt diff --git a/benchmarks/torchvision_ddp/main.py b/benchmarks/torchvision_ddp/main.py index 40de9169c..7548cbd84 100755 --- a/benchmarks/torchvision_ddp/main.py +++ b/benchmarks/torchvision_ddp/main.py @@ -17,9 +17,10 @@ import torchvision.transforms as transforms import torchvision.models as torchvision_models -from benchmate.metrics import BenchObserver, StopProgram -import torchcompat.core as accelerator +from benchmate.metrics import StopProgram +from benchmate.observer import BenchObserver from benchmate.dataloader import imagenet_dataloader, dataloader_arguments +import torchcompat.core as accelerator def ddp_setup(rank, world_size): diff --git a/benchmarks/torchvision_ddp/requirements.cuda.txt b/benchmarks/torchvision_ddp/requirements.cuda.txt index 385caea2c..104395a9f 100644 --- a/benchmarks/torchvision_ddp/requirements.cuda.txt +++ b/benchmarks/torchvision_ddp/requirements.cuda.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.cuda.txt .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision_ddp/requirements.in # ---extra-index-url https://download.pytorch.org/whl/cu118 +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera @@ -30,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -fsspec==2023.10.0 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch @@ -44,11 +39,11 @@ giving==0.4.2 # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch @@ -56,7 +51,7 @@ markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -68,31 +63,87 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -108,11 +159,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir @@ -120,34 +167,44 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -torch==2.1.0+cu118 +torch==2.3.1+cu121 # via - # -r benchmarks/torchvision/requirements.in + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat # torchvision -torchvision==0.16.0+cu118 - # via -r benchmarks/torchvision/requirements.in -tqdm==4.66.1 - # via -r benchmarks/torchvision/requirements.in -triton==2.1.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # torch -typing-extensions==4.8.0 + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving voir==0.2.15 - # via -r benchmarks/torchvision/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.hpu.txt b/benchmarks/torchvision_ddp/requirements.hpu.txt index 21560a4bd..7b53c2819 100644 --- a/benchmarks/torchvision_ddp/requirements.hpu.txt +++ b/benchmarks/torchvision_ddp/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/torchvision/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-torchvision.txt milabench/benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.hpu.txt .pin/tmp-constraints-hpu-torchvision.txt benchmarks/torchvision_ddp/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -109,7 +113,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -162,25 +166,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/torchvision/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat # torchvision -torchcompat==0.0.1 - # via -r milabench/benchmarks/torchvision/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/torchvision/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/torchvision/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -190,4 +203,7 @@ varname==0.10.0 # -c .pin/../.pin/constraints-hpu-torch.txt # giving voir==0.2.15 - # via -r milabench/benchmarks/torchvision/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.rocm.txt b/benchmarks/torchvision_ddp/requirements.rocm.txt index f9bb1d4c5..dc6f81cd3 100644 --- a/benchmarks/torchvision_ddp/requirements.rocm.txt +++ b/benchmarks/torchvision_ddp/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.rocm.txt .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision_ddp/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,19 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,11 +63,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,19 +75,23 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +99,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +111,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,31 +119,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via - # -r benchmarks/torchvision/requirements.in - # pytorch-triton-rocm + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/torchvision/requirements.in -tqdm==4.66.1 - # via -r benchmarks/torchvision/requirements.in -typing-extensions==4.8.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving voir==0.2.15 - # via -r benchmarks/torchvision/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.xpu.txt b/benchmarks/torchvision_ddp/requirements.xpu.txt index 0e529ce1f..3d3cf6603 100644 --- a/benchmarks/torchvision_ddp/requirements.xpu.txt +++ b/benchmarks/torchvision_ddp/requirements.xpu.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/torchvision/requirements.xpu.txt .pin/tmp-constraints-xpu-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.xpu.txt .pin/tmp-constraints-xpu-torchvision.txt benchmarks/torchvision_ddp/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera @@ -30,11 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -fsspec==2024.2.0 + # triton +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -43,11 +39,11 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -idna==3.7 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -jinja2==3.1.3 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -67,7 +63,7 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # sympy -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -75,6 +71,58 @@ numpy==1.26.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -87,11 +135,15 @@ pillow==10.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -107,10 +159,6 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # torchvision rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -119,34 +167,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt - # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat # torchvision -torchvision==0.16.0a0+cxx11.abi +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt - # -r benchmarks/torchvision/requirements.in -tqdm==4.66.2 + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # -r benchmarks/torchvision/requirements.in -typing-extensions==4.11.0 + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -155,4 +209,4 @@ voir==0.2.15 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt - # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmate/benchmate/metrics.py b/benchmate/benchmate/metrics.py index d46c500ab..975443e23 100644 --- a/benchmate/benchmate/metrics.py +++ b/benchmate/benchmate/metrics.py @@ -73,7 +73,7 @@ def materialize(self, *args, **kwargs): def push(self, pusher): """Iterate through data and push metrics.""" for args, kwargs in self.delayed: - pusher(self.materialize(*args, **kwargs)) + pusher(**self.materialize(*args, **kwargs)) self.delayed = [] diff --git a/milabench/_version.py b/milabench/_version.py index ced2a5852..3e7d71512 100644 --- a/milabench/_version.py +++ b/milabench/_version.py @@ -1,5 +1,5 @@ """This file is generated, do not modify""" -__tag__ = "v0.0.10-145-gc151b985" -__commit__ = "c151b98546f32d9c0671507f8526ed13598e3407" -__date__ = "2024-06-11 14:30:04 -0400" +__tag__ = "v0.0.10-147-gc6540c3e" +__commit__ = "c6540c3e470222e44b4a841954593185db49b111" +__date__ = "2024-06-12 07:11:39 -0400" diff --git a/milabench/report.py b/milabench/report.py index a51b1ac3a..5eacdad27 100644 --- a/milabench/report.py +++ b/milabench/report.py @@ -354,7 +354,7 @@ def fmtcol(col): _formatters = { "fail": "{:4.0f}".format, - "n": "{:.0f}".format, + "n": "{:3.0f}".format, "std": "{:10.2f}".format, "iqr": "{:10.2f}".format, "perf": "{:10.2f}".format, diff --git a/scripts/article/run_cuda.sh b/scripts/article/run_cuda.sh index 5a8f73e96..26e789da8 100644 --- a/scripts/article/run_cuda.sh +++ b/scripts/article/run_cuda.sh @@ -20,7 +20,7 @@ install_prepare() { fi if [ ! -d "$MILABENCH_WORDIR/milabench" ]; then - git clone https://github.com/mila-iqia/milabench.git -b intel + git clone https://github.com/mila-iqia/milabench.git fi . $MILABENCH_WORDIR/env/bin/activate @@ -60,8 +60,6 @@ fi cd $MILABENCH_WORDIR -(cd $MILABENCH_WORDIR/milabench && git pull origin intel) - # # Run the benchmakrs milabench run "$@" diff --git a/scripts/article/run_cuda_dev.sh b/scripts/article/run_cuda_dev.sh index 35faeb51f..0c73fefe3 100644 --- a/scripts/article/run_cuda_dev.sh +++ b/scripts/article/run_cuda_dev.sh @@ -18,8 +18,8 @@ install_prepare() { virtualenv $MILABENCH_WORDIR/env if [ ! -d "$MILABENCH_WORDIR/milabench" ]; then - git clone https://github.com/mila-iqia/milabench.git -b intel - git clone https://github.com/Delaunay/voir.git -b async_timer + git clone https://github.com/mila-iqia/milabench.git + git clone https://github.com/Delaunay/voir.git git clone https://github.com/Delaunay/torchcompat.git fi diff --git a/scripts/article/run_hpu.sh b/scripts/article/run_hpu.sh index f6add4850..ba90f5868 100644 --- a/scripts/article/run_hpu.sh +++ b/scripts/article/run_hpu.sh @@ -15,7 +15,7 @@ install_prepare() { virtualenv $MILABENCH_WORDIR/env - git clone https://github.com/mila-iqia/milabench.git -b intel + git clone https://github.com/mila-iqia/milabench.git git clone https://github.com/huggingface/optimum-habana.git wget -nv https://vault.habana.ai/artifactory/gaudi-installer/1.15.1/habanalabs-installer.sh diff --git a/scripts/article/run_rocm.sh b/scripts/article/run_rocm.sh index 819374e66..79e736c20 100644 --- a/scripts/article/run_rocm.sh +++ b/scripts/article/run_rocm.sh @@ -17,7 +17,7 @@ install_prepare() { virtualenv $MILABENCH_WORDIR/env - git clone https://github.com/mila-iqia/milabench.git -b intel + git clone https://github.com/mila-iqia/milabench.git . $MILABENCH_WORDIR/env/bin/activate pip install -e $MILABENCH_WORDIR/milabench diff --git a/scripts/article/run_xpu.sh b/scripts/article/run_xpu.sh index 86c741107..1effef43f 100644 --- a/scripts/article/run_xpu.sh +++ b/scripts/article/run_xpu.sh @@ -17,7 +17,7 @@ install_prepare() { virtualenv $MILABENCH_WORDIR/env - git clone https://github.com/mila-iqia/milabench.git -b intel + git clone https://github.com/mila-iqia/milabench.git # XPU manager is necessary wget -nv https://github.com/intel/xpumanager/releases/download/V1.2.36/xpumanager_1.2.36_20240428.081009.377f9162.u22.04_amd64.deb diff --git a/scripts/update_pins.sh b/scripts/update_pins.sh index 23b6ac841..5b8b56626 100644 --- a/scripts/update_pins.sh +++ b/scripts/update_pins.sh @@ -1,6 +1,9 @@ module load cuda/12.3.2 + +export MILABENCH_BASE=output/ + MILABENCH_GPU_ARCH=cuda milabench pin -c constraints/cuda.txt --config config/standard.yaml --from-scratch MILABENCH_GPU_ARCH=rocm milabench pin -c constraints/rocm.txt --config config/standard.yaml --from-scratch MILABENCH_GPU_ARCH=xpu milabench pin -c constraints/xpu.txt --config config/standard.yaml --from-scratch diff --git a/tests/test_summary/test_report.txt b/tests/test_summary/test_report.txt index 937a59561..a7039f7db 100644 --- a/tests/test_summary/test_report.txt +++ b/tests/test_summary/test_report.txt @@ -2,8 +2,8 @@ Source: XXX ================= Benchmark results ================= -bench | fail | n | perf | sem% | std% | peak_memory | score | weight -benchio | 0 | 4 | 7979.82 | 2.9% | 17.2% | -1 | 7979.82 | 2.00 +bench | fail | n | perf | sem% | std% | peak_memory | score | weight +benchio | 0 | 4 | 7979.82 | 2.9% | 17.2% | -1 | 7979.82 | 2.00 Scores ------ diff --git a/tests/test_summary/test_report_folder_does_average.txt b/tests/test_summary/test_report_folder_does_average.txt index 5abe96e68..b7ad71e7e 100644 --- a/tests/test_summary/test_report_folder_does_average.txt +++ b/tests/test_summary/test_report_folder_does_average.txt @@ -2,8 +2,8 @@ Source: XXX ================= Benchmark results ================= -bench | fail | n | perf | sem% | std% | peak_memory | score | weight -benchio | 0 | 6 | 7878.45 | 2.5% | 18.0% | 24456 | 7878.45 | 2.00 +bench | fail | n | perf | sem% | std% | peak_memory | score | weight +benchio | 0 | 6 | 7878.45 | 2.5% | 18.0% | 24456 | 7878.45 | 2.00 Scores ------