diff --git a/.pin/constraints-cuda-torch.txt b/.pin/constraints-cuda-torch.txt index 17b18a647..d7b074e3e 100644 --- a/.pin/constraints-cuda-torch.txt +++ b/.pin/constraints-cuda-torch.txt @@ -159,7 +159,7 @@ importlib-resources==6.4.0 # torchcompat itsdangerous==2.2.0 # via flask -jax[cuda12-cudnn89,cuda12_cudnn89]==0.4.28 +jax[cuda12]==0.4.28 # via # -r benchmarks/brax/requirements.in # brax @@ -169,6 +169,10 @@ jax[cuda12-cudnn89,cuda12_cudnn89]==0.4.28 # mujoco-mjx # optax # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via jax jaxlib==0.4.28+cuda12.cudnn89 # via # brax @@ -279,31 +283,48 @@ numpy==1.26.4 # trimesh nvidia-cublas-cu12==12.1.3.1 # via + # jax # nvidia-cudnn-cu12 # nvidia-cusolver-cu12 # torch nvidia-cuda-cupti-cu12==12.1.105 - # via torch + # via + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via jax nvidia-cuda-nvrtc-cu12==12.1.105 # via torch nvidia-cuda-runtime-cu12==12.1.105 - # via torch + # via + # jax + # torch nvidia-cudnn-cu12==8.9.2.26 - # via torch + # via + # jax + # torch nvidia-cufft-cu12==11.0.2.54 - # via torch + # via + # jax + # torch nvidia-curand-cu12==10.3.2.106 # via torch nvidia-cusolver-cu12==11.4.5.107 - # via torch + # via + # jax + # torch nvidia-cusparse-cu12==12.1.0.106 # via + # jax # nvidia-cusolver-cu12 # torch nvidia-nccl-cu12==2.20.5 - # via torch + # via + # jax + # torch nvidia-nvjitlink-cu12==12.5.40 # via + # jax # nvidia-cusolver-cu12 # nvidia-cusparse-cu12 nvidia-nvtx-cu12==12.1.105 diff --git a/.pin/constraints-hpu-torch.txt b/.pin/constraints-hpu-torch.txt new file mode 100644 index 000000000..20f5f2672 --- /dev/null +++ b/.pin/constraints-hpu-torch.txt @@ -0,0 +1,621 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=.pin/constraints-hpu-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint + # tensorboard +accelerate==0.31.0 + # via -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 + # via + # datasets + # fsspec +aiosignal==1.3.1 + # via aiohttp +antlr4-python3-runtime==4.9.3 + # via omegaconf +asttokens==2.4.1 + # via giving +attrs==23.2.0 + # via aiohttp +blinker==1.8.2 + # via flask +brax==0.10.5 + # via -r benchmarks/brax/requirements.in +certifi==2024.6.2 + # via requests +charset-normalizer==3.3.2 + # via requests +chex==0.1.86 + # via optax +click==8.1.7 + # via flask +cloudpickle==3.0.0 + # via gym +codefind==0.1.6 + # via ptera +contextlib2==21.6.0 + # via ml-collections +datasets==2.19.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/llama/requirements.in + # evaluate +deepspeed==0.14.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/rwkv/requirements.in +dill==0.3.8 + # via + # datasets + # evaluate + # multiprocess +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via dm-env +docker==7.1.0 + # via torchx +docstring-parser==0.8.1 + # via torchx +etils[epath,epy]==1.9.1 + # via + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +evaluate==0.4.2 + # via -r benchmarks/accelerate_opt/requirements.in +executing==1.2.0 + # via varname +fairscale==0.4.13 + # via -r benchmarks/llama/requirements.in +fbgemm-gpu==0.7.0 + # via torchrec +filelock==3.14.0 + # via + # datasets + # huggingface-hub + # torch + # torchx + # transformers + # triton +fire==0.6.0 + # via -r benchmarks/llama/requirements.in +flask==3.0.3 + # via + # brax + # flask-cors +flask-cors==4.0.1 + # via brax +flax==0.8.4 + # via brax +frozenlist==1.4.1 + # via + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # datasets + # etils + # evaluate + # huggingface-hub + # pytorch-lightning + # torch + # torchx +future==1.0.0 + # via -r benchmarks/dlrm/requirements.in +giving==0.4.2 + # via + # ptera + # voir +glfw==2.7.0 + # via mujoco +graphviz==0.20.3 + # via torchviz +grpcio==1.64.1 + # via + # brax + # tensorboard +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym +hjson==3.1.0 + # via deepspeed +huggingface-hub==0.23.3 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # datasets + # evaluate + # tokenizers + # transformers +idna==3.7 + # via + # requests + # yarl +importlib-metadata==7.1.0 + # via torchx +importlib-resources==6.4.0 + # via + # etils + # torchcompat +itsdangerous==2.2.0 + # via flask +jax[cuda12]==0.4.28 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via brax +jinja2==3.1.4 + # via + # brax + # flask + # torch +joblib==1.4.2 + # via scikit-learn +lightning-utilities==0.11.2 + # via + # pytorch-lightning + # torchmetrics +markdown==3.6 + # via tensorboard +markdown-it-py==3.0.0 + # via rich +markupsafe==2.1.5 + # via + # jinja2 + # werkzeug +mdurl==0.1.2 + # via markdown-it-py +ml-collections==0.1.1 + # via brax +ml-dtypes==0.4.0 + # via + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via sympy +msgpack==1.0.8 + # via + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via brax +multidict==6.0.5 + # via + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # datasets + # evaluate +mypy-extensions==1.0.0 + # via typing-inspect +nest-asyncio==1.6.0 + # via orbax-checkpoint +networkx==3.3 + # via torch +ninja==1.11.1.1 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed +numpy==1.26.4 + # via + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # accelerate + # brax + # chex + # datasets + # deepspeed + # dm-env + # evaluate + # fairscale + # fbgemm-gpu + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # onnx + # opencv-python + # opt-einsum + # optax + # orbax-checkpoint + # pandas + # pyarrow + # pytorch-lightning + # scikit-learn + # scipy + # tensorboard + # tensorboardx + # tensorstore + # torchmetrics + # torchvision + # transformers + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via torch +omegaconf==2.3.0 + # via voir +onnx==1.16.1 + # via -r benchmarks/dlrm/requirements.in +opencv-python==4.10.0.82 + # via -r benchmarks/super-slomo/requirements.in +opt-einsum==3.3.0 + # via jax +optax==0.2.2 + # via + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # brax + # flax +ovld==0.3.5 + # via voir +packaging==24.1 + # via + # accelerate + # datasets + # deepspeed + # evaluate + # huggingface-hub + # lightning-utilities + # pytorch-lightning + # tensorboardx + # torchmetrics + # transformers +pandas==2.2.2 + # via + # datasets + # evaluate +pillow==10.3.0 + # via + # brax + # torchvision +protobuf==4.25.3 + # via + # onnx + # orbax-checkpoint + # tensorboard + # tensorboardx +psutil==5.9.8 + # via + # accelerate + # deepspeed + # voir +ptera==1.4.1 + # via voir +py-cpuinfo==9.0.0 + # via deepspeed +pyarrow==16.1.0 + # via datasets +pyarrow-hotfix==0.6 + # via datasets +pydantic==1.10.15 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed +pydot==2.0.0 + # via -r benchmarks/dlrm/requirements.in +pygments==2.18.0 + # via rich +pynvml==11.5.0 + # via + # deepspeed + # voir +pyopengl==3.1.7 + # via mujoco +pyparsing==3.1.2 + # via pydot +pyre-extensions==0.0.30 + # via torchx +python-dateutil==2.9.0.post0 + # via pandas +pytinyrenderer==0.0.14 + # via brax +pytorch-lightning==1.9.5 + # via -r benchmarks/rwkv/requirements.in +pytz==2024.1 + # via pandas +pyyaml==6.0.1 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # datasets + # flax + # huggingface-hub + # ml-collections + # omegaconf + # orbax-checkpoint + # pytorch-lightning + # torchx + # transformers +reactivex==4.0.4 + # via giving +regex==2024.5.15 + # via transformers +requests==2.32.3 + # via + # datasets + # docker + # evaluate + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # flax + # voir +safetensors==0.4.3 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # transformers +scikit-learn==1.5.0 + # via -r benchmarks/dlrm/requirements.in +scipy==1.13.1 + # via + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx + # scikit-learn +sentencepiece==0.2.0 + # via -r benchmarks/llama/requirements.in +six==1.16.0 + # via + # asttokens + # fire + # ml-collections + # python-dateutil + # tensorboard +sympy==1.12.1 + # via torch +tabulate==0.9.0 + # via torchx +tensorboard==2.17.0 + # via -r benchmarks/dlrm/requirements.in +tensorboard-data-server==0.7.2 + # via tensorboard +tensorboardx==2.6.2.2 + # via brax +tensorstore==0.1.60 + # via + # flax + # orbax-checkpoint +termcolor==2.4.0 + # via fire +threadpoolctl==3.5.0 + # via scikit-learn +tokenizers==0.19.1 + # via transformers +toolz==0.12.1 + # via chex +torch==2.3.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # accelerate + # deepspeed + # fairscale + # pytorch-lightning + # torchaudio + # torchcompat + # torchmetrics + # torchvision + # torchviz +torchaudio==2.3.1 + # via -r benchmarks/accelerate_opt/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../constraints/hpu.txt + # -r benchmarks/flops/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +torchmetrics==1.0.3 + # via + # pytorch-lightning + # torchrec +torchrec==0.7.0 + # via -r benchmarks/dlrm/requirements.in +torchvision==0.18.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +torchviz==0.0.2 + # via -r benchmarks/dlrm/requirements.in +torchx==0.5.0 + # via -r benchmarks/dlrm/requirements.in +tqdm==4.66.4 + # via + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # datasets + # deepspeed + # evaluate + # huggingface-hub + # pytorch-lightning + # torchrec + # transformers +transformers==4.41.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in +trimesh==4.4.1 + # via + # brax + # mujoco-mjx +triton==2.3.1 + # via torch +typing-extensions==4.12.2 + # via + # brax + # chex + # etils + # flax + # huggingface-hub + # lightning-utilities + # orbax-checkpoint + # pydantic + # pyre-extensions + # pytorch-lightning + # reactivex + # torch + # typing-inspect +typing-inspect==0.9.0 + # via pyre-extensions +tzdata==2024.1 + # via pandas +urllib3==1.26.18 + # via + # docker + # requests + # torchx +varname==0.10.0 + # via giving +voir==0.2.14 + # via + # -c .pin/../constraints/hpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +werkzeug==3.0.3 + # via + # flask + # tensorboard +xxhash==3.4.1 + # via + # datasets + # evaluate +yarl==1.9.4 + # via aiohttp +zipp==3.19.2 + # via + # etils + # importlib-metadata + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/.pin/constraints-rocm-torch.txt b/.pin/constraints-rocm-torch.txt index a220c5c8d..f4e0c6336 100644 --- a/.pin/constraints-rocm-torch.txt +++ b/.pin/constraints-rocm-torch.txt @@ -2,15 +2,27 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=.pin/constraints-rocm-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in +# pip-compile --output-file=.pin/constraints-rocm-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -absl-py==2.0.0 - # via tensorboard -accelerate==0.24.1 +absl-py==2.1.0 + # via + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint + # tensorboard +accelerate==0.31.0 # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.6 +aiohttp==3.9.5 # via # datasets # fsspec @@ -20,83 +32,114 @@ antlr4-python3-runtime==4.9.3 # via omegaconf asttokens==2.4.1 # via giving -async-timeout==4.0.3 +attrs==23.2.0 # via aiohttp -attrs==23.1.0 - # via aiohttp -cachetools==5.3.2 - # via google-auth -certifi==2023.7.22 +blinker==1.8.2 + # via flask +brax==0.10.5 + # via -r benchmarks/brax/requirements.in +certifi==2024.6.2 # via requests charset-normalizer==3.3.2 - # via - # aiohttp - # requests -cmake==3.27.7 - # via pytorch-triton-rocm -codefind==0.1.3 + # via requests +chex==0.1.86 + # via optax +click==8.1.7 + # via flask +cloudpickle==3.0.0 + # via gym +codefind==0.1.6 # via ptera -datasets==2.14.6 +contextlib2==21.6.0 + # via ml-collections +datasets==2.19.2 # via # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/llama/requirements.in # evaluate -deepspeed==0.12.2 +deepspeed==0.14.2 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/rwkv/requirements.in -dill==0.3.7 +dill==0.3.8 # via # datasets # evaluate # multiprocess -docker==6.1.3 +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via dm-env +docker==7.1.0 # via torchx docstring-parser==0.8.1 # via torchx -evaluate==0.4.1 +etils[epath,epy]==1.9.1 + # via + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +evaluate==0.4.2 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via varname -fbgemm-gpu==0.5.0 +fairscale==0.4.13 + # via -r benchmarks/llama/requirements.in +fbgemm-gpu==0.7.0+rocm6.0 # via torchrec -filelock==3.13.1 +filelock==3.14.0 # via + # datasets # huggingface-hub # pytorch-triton-rocm # torch # torchx # transformers -frozenlist==1.4.0 +fire==0.6.0 + # via -r benchmarks/llama/requirements.in +flask==3.0.3 + # via + # brax + # flask-cors +flask-cors==4.0.1 + # via brax +flax==0.8.4 + # via brax +frozenlist==1.4.1 # via # aiohttp # aiosignal -fsspec[http]==2023.10.0 +fsspec[http]==2024.3.1 # via # datasets + # etils # evaluate # huggingface-hub # pytorch-lightning # torch # torchx -future==0.18.3 +future==1.0.0 # via -r benchmarks/dlrm/requirements.in giving==0.4.2 # via # ptera # voir -google-auth==2.23.4 +glfw==2.7.0 + # via mujoco +graphviz==0.20.3 + # via torchviz +grpcio==1.64.1 # via - # google-auth-oauthlib + # brax # tensorboard -google-auth-oauthlib==1.1.0 - # via tensorboard -graphviz==0.20.1 - # via torchviz -grpcio==1.59.2 - # via tensorboard +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym hjson==3.1.0 # via deepspeed -huggingface-hub==0.17.3 +huggingface-hub==0.23.3 # via # -r benchmarks/timm/requirements.in # accelerate @@ -104,235 +147,388 @@ huggingface-hub==0.17.3 # evaluate # tokenizers # transformers -idna==3.4 +idna==3.7 # via # requests # yarl -importlib-metadata==6.8.0 +importlib-metadata==7.1.0 # via torchx -jinja2==3.1.2 - # via torch -joblib==1.3.2 +importlib-resources==6.4.0 + # via + # etils + # torchcompat +itsdangerous==2.2.0 + # via flask +jax[cuda12]==0.4.28 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via brax +jinja2==3.1.4 + # via + # brax + # flask + # torch +joblib==1.4.2 # via scikit-learn -lightning-utilities==0.9.0 +lightning-utilities==0.11.2 # via # pytorch-lightning # torchmetrics -lit==17.0.4 - # via pytorch-triton-rocm -markdown==3.5.1 +markdown==3.6 # via tensorboard markdown-it-py==3.0.0 # via rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # jinja2 # werkzeug mdurl==0.1.2 # via markdown-it-py +ml-collections==0.1.1 + # via brax +ml-dtypes==0.4.0 + # via + # jax + # jaxlib + # tensorstore mpmath==1.3.0 # via sympy -multidict==6.0.4 +msgpack==1.0.8 + # via + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via brax +multidict==6.0.5 # via # aiohttp # yarl -multiprocess==0.70.15 +multiprocess==0.70.16 # via # datasets # evaluate mypy-extensions==1.0.0 # via typing-inspect -networkx==3.2.1 +nest-asyncio==1.6.0 + # via orbax-checkpoint +networkx==3.3 # via torch ninja==1.11.1.1 # via # -r benchmarks/rwkv/requirements.in # deepspeed -numpy==1.26.1 +numpy==1.26.4 # via # -r benchmarks/dlrm/requirements.in # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in # accelerate + # brax + # chex # datasets # deepspeed + # dm-env # evaluate + # fairscale # fbgemm-gpu + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco # onnx # opencv-python + # opt-einsum + # optax + # orbax-checkpoint # pandas # pyarrow # pytorch-lightning # scikit-learn # scipy # tensorboard + # tensorboardx + # tensorstore # torchmetrics # torchvision # transformers -oauthlib==3.2.2 - # via requests-oauthlib + # trimesh +nvidia-cublas-cu12==12.5.2.13 + # via + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 +nvidia-cuda-cupti-cu12==12.5.39 + # via jax +nvidia-cuda-nvcc-cu12==12.5.40 + # via jax +nvidia-cuda-nvrtc-cu12==12.5.40 + # via nvidia-cudnn-cu12 +nvidia-cuda-runtime-cu12==12.5.39 + # via jax +nvidia-cudnn-cu12==8.9.7.29 + # via jax +nvidia-cufft-cu12==11.2.3.18 + # via jax +nvidia-cusolver-cu12==11.6.2.40 + # via jax +nvidia-cusparse-cu12==12.4.1.24 + # via + # jax + # nvidia-cusolver-cu12 +nvidia-nccl-cu12==2.21.5 + # via jax +nvidia-nvjitlink-cu12==12.5.40 + # via + # jax + # nvidia-cufft-cu12 + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 omegaconf==2.3.0 # via voir -onnx==1.15.0 +onnx==1.16.1 # via -r benchmarks/dlrm/requirements.in -opencv-python==4.8.1.78 +opencv-python==4.10.0.82 # via -r benchmarks/super-slomo/requirements.in -ovld==0.3.2 +opt-einsum==3.3.0 + # via jax +optax==0.2.2 + # via + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # brax + # flax +ovld==0.3.5 # via voir -packaging==23.2 +packaging==24.1 # via # accelerate # datasets # deepspeed - # docker # evaluate # huggingface-hub # lightning-utilities # pytorch-lightning + # tensorboardx # torchmetrics # transformers -pandas==2.1.2 +pandas==2.2.2 # via # datasets # evaluate -pillow==10.1.0 - # via torchvision -protobuf==4.23.4 +pillow==10.3.0 + # via + # brax + # torchvision +protobuf==4.25.3 # via # onnx + # orbax-checkpoint # tensorboard -psutil==5.9.6 + # tensorboardx +psutil==5.9.8 # via # accelerate # deepspeed + # voir ptera==1.4.1 # via voir py-cpuinfo==9.0.0 # via deepspeed -pyarrow==14.0.0 +pyarrow==16.1.0 # via datasets -pyasn1==0.5.0 - # via - # pyasn1-modules - # rsa -pyasn1-modules==0.3.0 - # via google-auth -pydantic==1.10.13 +pyarrow-hotfix==0.6 + # via datasets +pydantic==1.10.15 # via # -r benchmarks/rwkv/requirements.in # deepspeed -pydot==1.4.2 +pydot==2.0.0 # via -r benchmarks/dlrm/requirements.in -pygments==2.16.1 +pygments==2.18.0 # via rich pynvml==11.5.0 # via # deepspeed # voir -pyparsing==3.1.1 +pyopengl==3.1.7 + # via mujoco +pyparsing==3.1.2 # via pydot pyre-extensions==0.0.30 # via torchx -python-dateutil==2.8.2 +python-dateutil==2.9.0.post0 # via pandas +pytinyrenderer==0.0.14 + # via brax pytorch-lightning==1.9.5 # via -r benchmarks/rwkv/requirements.in -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via torch -pytz==2023.3.post1 +pytz==2024.1 # via pandas pyyaml==6.0.1 # via # -r benchmarks/timm/requirements.in # accelerate # datasets + # flax # huggingface-hub + # ml-collections # omegaconf + # orbax-checkpoint # pytorch-lightning # torchx # transformers reactivex==4.0.4 # via giving -regex==2023.10.3 +regex==2024.5.15 # via transformers -requests==2.31.0 +requests==2.32.3 # via # datasets # docker # evaluate - # fsspec # huggingface-hub - # requests-oauthlib - # responses - # tensorboard - # torchvision # transformers -requests-oauthlib==1.3.1 - # via google-auth-oauthlib -responses==0.18.0 - # via evaluate -rich==13.6.0 +rich==13.7.1 # via # -r benchmarks/accelerate_opt/requirements.in + # flax # voir -rsa==4.9 - # via google-auth -safetensors==0.4.0 +safetensors==0.4.3 # via # -r benchmarks/timm/requirements.in + # accelerate # transformers -scikit-learn==1.3.2 +scikit-learn==1.5.0 # via -r benchmarks/dlrm/requirements.in -scipy==1.11.3 - # via scikit-learn +scipy==1.13.1 + # via + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx + # scikit-learn +sentencepiece==0.2.0 + # via -r benchmarks/llama/requirements.in six==1.16.0 # via # asttokens + # fire + # ml-collections # python-dateutil # tensorboard -sympy==1.12 +sympy==1.12.1 # via torch tabulate==0.9.0 # via torchx -tensorboard==2.15.1 +tensorboard==2.17.0 # via -r benchmarks/dlrm/requirements.in tensorboard-data-server==0.7.2 # via tensorboard -threadpoolctl==3.2.0 +tensorboardx==2.6.2.2 + # via brax +tensorstore==0.1.60 + # via + # flax + # orbax-checkpoint +termcolor==2.4.0 + # via fire +threadpoolctl==3.5.0 # via scikit-learn -tokenizers==0.14.1 +tokenizers==0.19.1 # via transformers -torch==2.1.0+rocm5.6 +toolz==0.12.1 + # via chex +torch==2.3.1+rocm6.0 # via # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in # accelerate # deepspeed + # fairscale # pytorch-lightning - # pytorch-triton-rocm # torchaudio + # torchcompat # torchmetrics # torchvision # torchviz -torchaudio==2.1.0+rocm5.6 +torchaudio==2.3.1+rocm6.0 # via -r benchmarks/accelerate_opt/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../constraints/rocm.txt + # -r benchmarks/flops/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in torchmetrics==1.0.3 # via # pytorch-lightning # torchrec -torchrec==0.5.0 +torchrec==0.7.0 # via -r benchmarks/dlrm/requirements.in -torchvision==0.16.0+rocm5.6 +torchvision==0.18.1+rocm6.0 # via # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/flops/requirements.in # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in torchx==0.5.0 # via -r benchmarks/dlrm/requirements.in -tqdm==4.66.1 +tqdm==4.66.4 # via + # -r benchmarks/dlrm/requirements.in # -r benchmarks/flops/requirements.in + # -r benchmarks/super-slomo/requirements.in # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in # datasets # deepspeed # evaluate @@ -340,14 +536,24 @@ tqdm==4.66.1 # pytorch-lightning # torchrec # transformers -transformers==4.35.0 +transformers==4.41.2 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/huggingface/requirements.in -typing-extensions==4.8.0 + # -r benchmarks/llama/requirements.in +trimesh==4.4.1 + # via + # brax + # mujoco-mjx +typing-extensions==4.12.2 # via + # brax + # chex + # etils + # flax # huggingface-hub # lightning-utilities + # orbax-checkpoint # pydantic # pyre-extensions # pytorch-lightning @@ -356,32 +562,44 @@ typing-extensions==4.8.0 # typing-inspect typing-inspect==0.9.0 # via pyre-extensions -tzdata==2023.3 +tzdata==2024.1 # via pandas urllib3==1.26.18 # via # docker # requests - # responses # torchx varname==0.10.0 # via giving -voir==0.2.11 +voir==0.2.14 # via + # -c .pin/../constraints/rocm.txt # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in # -r benchmarks/rwkv/requirements.in -websocket-client==1.6.4 - # via docker -werkzeug==3.0.1 - # via tensorboard + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +werkzeug==3.0.3 + # via + # flask + # tensorboard xxhash==3.4.1 # via # datasets # evaluate -yarl==1.9.2 +yarl==1.9.4 # via aiohttp -zipp==3.17.0 - # via importlib-metadata +zipp==3.19.2 + # via + # etils + # importlib-metadata # The following packages are considered to be unsafe in a requirements file: # setuptools diff --git a/.pin/constraints-xpu-torch.txt b/.pin/constraints-xpu-torch.txt new file mode 100644 index 000000000..09b85c920 --- /dev/null +++ b/.pin/constraints-xpu-torch.txt @@ -0,0 +1,626 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=.pin/constraints-xpu-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint + # tensorboard +accelerate==0.31.0 + # via -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 + # via + # datasets + # fsspec +aiosignal==1.3.1 + # via aiohttp +antlr4-python3-runtime==4.9.3 + # via omegaconf +asttokens==2.4.1 + # via giving +attrs==23.2.0 + # via aiohttp +blinker==1.8.2 + # via flask +brax==0.10.5 + # via -r benchmarks/brax/requirements.in +certifi==2024.6.2 + # via requests +charset-normalizer==3.3.2 + # via requests +chex==0.1.86 + # via optax +click==8.1.7 + # via flask +cloudpickle==3.0.0 + # via gym +codefind==0.1.6 + # via ptera +contextlib2==21.6.0 + # via ml-collections +datasets==2.19.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/llama/requirements.in + # evaluate +deepspeed==0.14.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/rwkv/requirements.in +dill==0.3.8 + # via + # datasets + # evaluate + # multiprocess +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via dm-env +docker==7.1.0 + # via torchx +docstring-parser==0.8.1 + # via torchx +etils[epath,epy]==1.9.1 + # via + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +evaluate==0.4.2 + # via -r benchmarks/accelerate_opt/requirements.in +executing==1.2.0 + # via varname +fairscale==0.4.13 + # via -r benchmarks/llama/requirements.in +fbgemm-gpu==0.7.0 + # via torchrec +filelock==3.14.0 + # via + # datasets + # huggingface-hub + # torch + # torchx + # transformers + # triton +fire==0.6.0 + # via -r benchmarks/llama/requirements.in +flask==3.0.3 + # via + # brax + # flask-cors +flask-cors==4.0.1 + # via brax +flax==0.8.4 + # via brax +frozenlist==1.4.1 + # via + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # datasets + # etils + # evaluate + # huggingface-hub + # pytorch-lightning + # torch + # torchx +future==1.0.0 + # via -r benchmarks/dlrm/requirements.in +giving==0.4.2 + # via + # ptera + # voir +glfw==2.7.0 + # via mujoco +graphviz==0.20.3 + # via torchviz +grpcio==1.64.1 + # via + # brax + # tensorboard +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym +hjson==3.1.0 + # via deepspeed +huggingface-hub==0.23.3 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # datasets + # evaluate + # tokenizers + # transformers +idna==3.7 + # via + # requests + # yarl +importlib-metadata==7.1.0 + # via torchx +importlib-resources==6.4.0 + # via + # etils + # torchcompat +itsdangerous==2.2.0 + # via flask +jax[cuda12]==0.4.28 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via brax +jinja2==3.1.4 + # via + # brax + # flask + # torch +joblib==1.4.2 + # via scikit-learn +lightning-utilities==0.11.2 + # via + # pytorch-lightning + # torchmetrics +markdown==3.6 + # via tensorboard +markdown-it-py==3.0.0 + # via rich +markupsafe==2.1.5 + # via + # jinja2 + # werkzeug +mdurl==0.1.2 + # via markdown-it-py +ml-collections==0.1.1 + # via brax +ml-dtypes==0.4.0 + # via + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via sympy +msgpack==1.0.8 + # via + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via brax +multidict==6.0.5 + # via + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # datasets + # evaluate +mypy-extensions==1.0.0 + # via typing-inspect +nest-asyncio==1.6.0 + # via orbax-checkpoint +networkx==3.3 + # via torch +ninja==1.11.1.1 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed +numpy==1.26.4 + # via + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # accelerate + # brax + # chex + # datasets + # deepspeed + # dm-env + # evaluate + # fairscale + # fbgemm-gpu + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # onnx + # opencv-python + # opt-einsum + # optax + # orbax-checkpoint + # pandas + # pyarrow + # pytorch-lightning + # scikit-learn + # scipy + # tensorboard + # tensorboardx + # tensorstore + # torchmetrics + # torchvision + # transformers + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via torch +omegaconf==2.3.0 + # via voir +onnx==1.16.1 + # via -r benchmarks/dlrm/requirements.in +opencv-python==4.10.0.82 + # via -r benchmarks/super-slomo/requirements.in +opt-einsum==3.3.0 + # via jax +optax==0.2.2 + # via + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # brax + # flax +ovld==0.3.5 + # via voir +packaging==24.1 + # via + # accelerate + # datasets + # deepspeed + # evaluate + # huggingface-hub + # lightning-utilities + # pytorch-lightning + # tensorboardx + # torchmetrics + # transformers +pandas==2.2.2 + # via + # datasets + # evaluate +pillow==10.3.0 + # via + # brax + # torchvision +protobuf==4.25.3 + # via + # onnx + # orbax-checkpoint + # tensorboard + # tensorboardx +psutil==5.9.8 + # via + # accelerate + # deepspeed + # voir +ptera==1.4.1 + # via voir +py-cpuinfo==9.0.0 + # via deepspeed +pyarrow==16.1.0 + # via datasets +pyarrow-hotfix==0.6 + # via datasets +pydantic==1.10.15 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed +pydot==2.0.0 + # via -r benchmarks/dlrm/requirements.in +pygments==2.18.0 + # via rich +pynvml==11.5.0 + # via + # deepspeed + # voir +pyopengl==3.1.7 + # via mujoco +pyparsing==3.1.2 + # via pydot +pyre-extensions==0.0.30 + # via torchx +python-dateutil==2.9.0.post0 + # via pandas +pytinyrenderer==0.0.14 + # via brax +pytorch-lightning==1.9.5 + # via -r benchmarks/rwkv/requirements.in +pytz==2024.1 + # via pandas +pyyaml==6.0.1 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # datasets + # flax + # huggingface-hub + # ml-collections + # omegaconf + # orbax-checkpoint + # pytorch-lightning + # torchx + # transformers +reactivex==4.0.4 + # via giving +regex==2024.5.15 + # via transformers +requests==2.32.3 + # via + # datasets + # docker + # evaluate + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # flax + # voir +safetensors==0.4.3 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # transformers +scikit-learn==1.5.0 + # via -r benchmarks/dlrm/requirements.in +scipy==1.13.1 + # via + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx + # scikit-learn +sentencepiece==0.2.0 + # via -r benchmarks/llama/requirements.in +six==1.16.0 + # via + # asttokens + # fire + # ml-collections + # python-dateutil + # tensorboard +sympy==1.12.1 + # via torch +tabulate==0.9.0 + # via torchx +tensorboard==2.17.0 + # via -r benchmarks/dlrm/requirements.in +tensorboard-data-server==0.7.2 + # via tensorboard +tensorboardx==2.6.2.2 + # via brax +tensorstore==0.1.60 + # via + # flax + # orbax-checkpoint +termcolor==2.4.0 + # via fire +threadpoolctl==3.5.0 + # via scikit-learn +tokenizers==0.19.1 + # via transformers +toolz==0.12.1 + # via chex +torch==2.3.1 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # accelerate + # deepspeed + # fairscale + # pytorch-lightning + # torchaudio + # torchcompat + # torchmetrics + # torchvision + # torchviz +torchaudio==2.3.1 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/flops/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +torchmetrics==1.0.3 + # via + # pytorch-lightning + # torchrec +torchrec==0.7.0 + # via -r benchmarks/dlrm/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +torchviz==0.0.2 + # via -r benchmarks/dlrm/requirements.in +torchx==0.5.0 + # via -r benchmarks/dlrm/requirements.in +tqdm==4.66.4 + # via + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # datasets + # deepspeed + # evaluate + # huggingface-hub + # pytorch-lightning + # torchrec + # transformers +transformers==4.41.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in +trimesh==4.4.1 + # via + # brax + # mujoco-mjx +triton==2.3.1 + # via torch +typing-extensions==4.12.2 + # via + # brax + # chex + # etils + # flax + # huggingface-hub + # lightning-utilities + # orbax-checkpoint + # pydantic + # pyre-extensions + # pytorch-lightning + # reactivex + # torch + # typing-inspect +typing-inspect==0.9.0 + # via pyre-extensions +tzdata==2024.1 + # via pandas +urllib3==1.26.18 + # via + # docker + # requests + # torchx +varname==0.10.0 + # via giving +voir==0.2.14 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +werkzeug==3.0.3 + # via + # flask + # tensorboard +xxhash==3.4.1 + # via + # datasets + # evaluate +yarl==1.9.4 + # via aiohttp +zipp==3.19.2 + # via + # etils + # importlib-metadata + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/accelerate_opt/requirements.cuda.txt b/benchmarks/accelerate_opt/requirements.cuda.txt new file mode 100644 index 000000000..50f4a6514 --- /dev/null +++ b/benchmarks/accelerate_opt/requirements.cuda.txt @@ -0,0 +1,394 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.cuda.txt .pin/tmp-constraints-cuda-opt.txt benchmarks/accelerate_opt/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +accelerate==0.31.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # evaluate +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +dill==0.3.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate + # multiprocess +evaluate==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # torch + # transformers + # triton +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +hjson==3.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # datasets + # evaluate + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +ninja==1.11.1.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # datasets + # deepspeed + # evaluate + # pandas + # pyarrow + # torchvision + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # datasets + # deepspeed + # evaluate + # huggingface-hub + # transformers +pandas==2.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # deepspeed + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +py-cpuinfo==9.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +pyarrow==16.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +pydantic==1.10.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # voir +python-dateutil==2.9.0.post0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +pytz==2024.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # datasets + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # transformers +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens + # python-dateutil +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # accelerate + # deepspeed + # torchaudio + # torchvision +torchaudio==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # deepspeed + # evaluate + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # pydantic + # reactivex + # torch +tzdata==2024.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/accelerate_opt/requirements.in +xxhash==3.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp diff --git a/benchmarks/accelerate_opt/requirements.hpu.txt b/benchmarks/accelerate_opt/requirements.hpu.txt index a480dcd78..e1bad188f 100644 --- a/benchmarks/accelerate_opt/requirements.hpu.txt +++ b/benchmarks/accelerate_opt/requirements.hpu.txt @@ -1,11 +1,17 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/accelerate_opt/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-opt.txt milabench/benchmarks/accelerate_opt/requirements.in +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.hpu.txt .pin/tmp-constraints-hpu-opt.txt benchmarks/accelerate_opt/requirements.in # -accelerate==0.30.0 - # via -r milabench/benchmarks/accelerate_opt/requirements.in +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +accelerate==0.31.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -23,15 +29,11 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -async-timeout==4.0.3 - # via - # -c .pin/../.pin/constraints-hpu-torch.txt - # aiohttp attrs==23.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # aiohttp -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -39,16 +41,19 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera -datasets==2.19.1 +datasets==2.19.2 # via - # -r milabench/benchmarks/accelerate_opt/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in # evaluate deepspeed==0.14.2 - # via -r milabench/benchmarks/accelerate_opt/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in dill==0.3.8 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -56,7 +61,9 @@ dill==0.3.8 # evaluate # multiprocess evaluate==0.4.2 - # via -r milabench/benchmarks/accelerate_opt/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -90,7 +97,7 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed -huggingface-hub==0.23.0 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # accelerate @@ -195,7 +202,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -212,7 +219,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # accelerate @@ -244,7 +251,7 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed -pyarrow==16.0.0 +pyarrow==16.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -285,11 +292,11 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -regex==2024.4.28 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -298,7 +305,8 @@ requests==2.31.0 # transformers rich==13.7.1 # via - # -r milabench/benchmarks/accelerate_opt/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in # voir safetensors==0.4.3 # via @@ -310,7 +318,7 @@ six==1.16.0 # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens # python-dateutil -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch @@ -318,17 +326,22 @@ tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/accelerate_opt/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in # accelerate # deepspeed # torchaudio # torchvision -torchaudio==2.3.0 - # via -r milabench/benchmarks/accelerate_opt/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/accelerate_opt/requirements.in +torchaudio==2.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in tqdm==4.66.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -337,13 +350,15 @@ tqdm==4.66.4 # evaluate # huggingface-hub # transformers -transformers==4.40.2 - # via -r milabench/benchmarks/accelerate_opt/requirements.in -triton==2.3.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -362,8 +377,11 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/accelerate_opt/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/accelerate_opt/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index b56ff798f..72e0009f5 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -2,13 +2,18 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/accelerate_opt/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-opt.txt benchmarks/accelerate_opt/requirements.in +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.rocm.txt .pin/tmp-constraints-rocm-opt.txt benchmarks/accelerate_opt/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -accelerate==0.24.1 - # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.6 +accelerate==0.31.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets @@ -25,62 +30,59 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -async-timeout==4.0.3 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # aiohttp -attrs==23.1.0 +attrs==23.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp -certifi==2023.7.22 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # aiohttp # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera -datasets==2.14.6 +datasets==2.19.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.12.2 - # via -r benchmarks/accelerate_opt/requirements.in -dill==0.3.7 +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +dill==0.3.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate # multiprocess -evaluate==0.4.1 - # via -r benchmarks/accelerate_opt/requirements.in +evaluate==0.4.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets # huggingface-hub # pytorch-triton-rocm # torch # transformers -frozenlist==1.4.0 +frozenlist==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.10.0 +fsspec[http]==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets @@ -96,7 +98,7 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -huggingface-hub==0.17.3 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -104,24 +106,20 @@ huggingface-hub==0.17.3 # evaluate # tokenizers # transformers -idna==3.4 +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests # yarl -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -133,17 +131,17 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -multidict==6.0.4 +multidict==6.0.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # yarl -multiprocess==0.70.15 +multiprocess==0.70.16 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -151,7 +149,7 @@ ninja==1.11.1.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -166,11 +164,11 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -179,20 +177,21 @@ packaging==23.2 # evaluate # huggingface-hub # transformers -pandas==2.1.2 +pandas==2.2.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision -psutil==5.9.6 +psutil==5.9.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate # deepspeed + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -201,15 +200,19 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pyarrow==14.0.0 +pyarrow==16.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets -pydantic==1.10.13 +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +pydantic==1.10.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -218,15 +221,15 @@ pynvml==11.5.0 # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed # voir -python-dateutil==2.8.2 +python-dateutil==2.9.0.post0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pytz==2023.3.post1 +pytz==2024.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas @@ -242,58 +245,57 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -regex==2023.10.3 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate - # fsspec # huggingface-hub - # responses - # torchvision # transformers -responses==0.18.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # evaluate -rich==13.6.0 - # via # -r benchmarks/accelerate_opt/requirements.in # voir -safetensors==0.4.0 +safetensors==0.4.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # accelerate # transformers six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens # python-dateutil -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -tokenizers==0.14.1 +tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/accelerate_opt/requirements.in # accelerate # deepspeed - # pytorch-triton-rocm # torchaudio # torchvision -torchaudio==2.1.0+rocm5.6 - # via -r benchmarks/accelerate_opt/requirements.in -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/accelerate_opt/requirements.in -tqdm==4.66.1 +torchaudio==2.3.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets @@ -301,16 +303,18 @@ tqdm==4.66.1 # evaluate # huggingface-hub # transformers -transformers==4.35.0 - # via -r benchmarks/accelerate_opt/requirements.in -typing-extensions==4.8.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pydantic # reactivex # torch -tzdata==2023.3 +tzdata==2024.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas @@ -318,19 +322,21 @@ urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests - # responses varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/accelerate_opt/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/accelerate_opt/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate -yarl==1.9.2 +yarl==1.9.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp diff --git a/benchmarks/accelerate_opt/requirements.xpu.txt b/benchmarks/accelerate_opt/requirements.xpu.txt new file mode 100644 index 000000000..2b661a191 --- /dev/null +++ b/benchmarks/accelerate_opt/requirements.xpu.txt @@ -0,0 +1,397 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.xpu.txt .pin/tmp-constraints-xpu-opt.txt benchmarks/accelerate_opt/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +accelerate==0.31.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # evaluate +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +dill==0.3.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate + # multiprocess +evaluate==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # huggingface-hub + # torch + # transformers + # triton +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +hjson==3.1.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # datasets + # evaluate + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +ninja==1.11.1.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # datasets + # deepspeed + # evaluate + # pandas + # pyarrow + # torchvision + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # datasets + # deepspeed + # evaluate + # huggingface-hub + # transformers +pandas==2.2.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # deepspeed + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +py-cpuinfo==9.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +pyarrow==16.1.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets +pydantic==1.10.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # voir +python-dateutil==2.9.0.post0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pandas +pytz==2024.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pandas +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # datasets + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # transformers +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens + # python-dateutil +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # accelerate + # deepspeed + # torchaudio + # torchvision +torchaudio==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # deepspeed + # evaluate + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # pydantic + # reactivex + # torch +tzdata==2024.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pandas +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in +xxhash==3.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp diff --git a/benchmarks/brax/requirements.cuda.txt b/benchmarks/brax/requirements.cuda.txt new file mode 100644 index 000000000..5e7dc7c3d --- /dev/null +++ b/benchmarks/brax/requirements.cuda.txt @@ -0,0 +1,447 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/brax/requirements.cuda.txt .pin/tmp-constraints-cuda-brax.txt benchmarks/brax/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +blinker==1.8.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +brax==0.10.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/brax/requirements.in +chex==0.1.86 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # optax +click==8.1.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +cloudpickle==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # gym +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +contextlib2==21.6.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ml-collections +dm-env==1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +dm-tree==0.1.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # dm-env +etils[epath,epy]==1.9.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +flask==3.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # flask-cors +flask-cors==4.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +flax==0.8.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # etils + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +glfw==2.7.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # mujoco +grpcio==1.64.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +gym==0.26.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +gym-notices==0.0.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # gym +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # etils +itsdangerous==2.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +jax[cuda12]==0.4.28 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # flask + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +ml-collections==0.1.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +ml-dtypes==0.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +msgpack==1.0.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +nest-asyncio==1.6.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # orbax-checkpoint +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # chex + # dm-env + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # opt-einsum + # optax + # orbax-checkpoint + # scipy + # tensorboardx + # tensorstore + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +opt-einsum==3.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax +optax==0.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # flax +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboardx +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +protobuf==4.25.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # orbax-checkpoint + # tensorboardx +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyopengl==3.1.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # mujoco +pytinyrenderer==0.0.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax + # ml-collections + # omegaconf + # orbax-checkpoint +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax + # voir +scipy==1.13.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens + # ml-collections +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tensorboardx==2.6.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +tensorstore==0.1.60 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax + # orbax-checkpoint +toolz==0.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # chex +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/brax/requirements.in +trimesh==4.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # mujoco-mjx +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # chex + # etils + # flax + # orbax-checkpoint + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/brax/requirements.in +werkzeug==3.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +zipp==3.19.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # etils diff --git a/benchmarks/brax/requirements.hpu.txt b/benchmarks/brax/requirements.hpu.txt new file mode 100644 index 000000000..ed3084061 --- /dev/null +++ b/benchmarks/brax/requirements.hpu.txt @@ -0,0 +1,446 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/brax/requirements.hpu.txt .pin/tmp-constraints-hpu-brax.txt benchmarks/brax/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # giving +blinker==1.8.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flask +brax==0.10.5 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/brax/requirements.in +chex==0.1.86 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # optax +click==8.1.7 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flask +cloudpickle==3.0.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # gym +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # ptera +contextlib2==21.6.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # ml-collections +dm-env==1.6 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +dm-tree==0.1.8 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # dm-env +etils[epath,epy]==1.9.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch + # triton +flask==3.0.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # flask-cors +flask-cors==4.0.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +flax==0.8.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # etils + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # ptera + # voir +glfw==2.7.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # mujoco +grpcio==1.64.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +gym==0.26.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +gym-notices==0.0.8 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # gym +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # etils +itsdangerous==2.2.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flask +jax[cuda12]==0.4.28 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # flask + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # markdown-it-py +ml-collections==0.1.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +ml-dtypes==0.4.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # sympy +msgpack==1.0.8 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +nest-asyncio==1.6.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # orbax-checkpoint +networkx==3.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # chex + # dm-env + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # opt-einsum + # optax + # orbax-checkpoint + # scipy + # tensorboardx + # tensorstore + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +opt-einsum==3.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax +optax==0.2.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # flax +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # tensorboardx +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +protobuf==4.25.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # orbax-checkpoint + # tensorboardx +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +pyopengl==3.1.7 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # mujoco +pytinyrenderer==0.0.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flax + # ml-collections + # omegaconf + # orbax-checkpoint +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flax + # voir +scipy==1.13.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx +six==1.16.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # asttokens + # ml-collections +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +tensorboardx==2.6.2.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +tensorstore==0.1.60 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flax + # orbax-checkpoint +toolz==0.12.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # chex +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/brax/requirements.in +trimesh==4.4.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # mujoco-mjx +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # chex + # etils + # flax + # orbax-checkpoint + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/brax/requirements.in +werkzeug==3.0.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flask +zipp==3.19.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # etils diff --git a/benchmarks/brax/requirements.in b/benchmarks/brax/requirements.in index 35aef7be0..9db61b9e8 100644 --- a/benchmarks/brax/requirements.in +++ b/benchmarks/brax/requirements.in @@ -1,4 +1,4 @@ -jax[cuda12_cudnn89] +jax[cuda12] --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html torch brax diff --git a/benchmarks/brax/requirements.rocm.txt b/benchmarks/brax/requirements.rocm.txt index 118c6cff8..6e1503248 100644 --- a/benchmarks/brax/requirements.rocm.txt +++ b/benchmarks/brax/requirements.rocm.txt @@ -1,59 +1,61 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/brax/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-brax.txt benchmarks/brax/requirements.in +# pip-compile --output-file=benchmarks/brax/requirements.rocm.txt .pin/tmp-constraints-rocm-brax.txt benchmarks/brax/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -absl-py==1.4.0 +absl-py==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # chex # dm-env + # ml-collections # mujoco + # mujoco-mjx # optax # orbax-checkpoint antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -blinker==1.6.2 +blinker==1.8.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -brax==0.9.0 - # via -r benchmarks/brax/requirements.in -cached-property==1.5.2 +brax==0.10.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # orbax-checkpoint -chex==0.1.7 + # -r benchmarks/brax/requirements.in +chex==0.1.86 # via # -c .pin/../.pin/constraints-rocm-torch.txt # optax -click==8.1.3 +click==8.1.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -cloudpickle==2.2.1 +cloudpickle==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # gym -cmake==3.26.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 + # ptera +contextlib2==21.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # ptera + # ml-collections dm-env==1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -61,45 +63,51 @@ dm-env==1.6 dm-tree==0.1.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # chex # dm-env -etils==1.3.0 +etils[epath,epy]==1.9.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax + # mujoco + # mujoco-mjx # orbax-checkpoint executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -flask==2.3.2 +flask==3.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # flask-cors -flask-cors==3.0.10 +flask-cors==4.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -flax==0.6.10 +flax==0.8.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # etils + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -glfw==2.5.9 +glfw==2.7.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # mujoco -grpcio==1.54.2 +grpcio==1.64.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax @@ -111,56 +119,58 @@ gym-notices==0.0.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # gym -importlib-metadata==6.6.0 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # flask - # gym -importlib-resources==5.12.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # orbax-checkpoint -itsdangerous==2.1.2 + # etils +itsdangerous==2.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -jax[cuda11_cudnn82]==0.4.8 +jax[cuda12]==0.4.28 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/brax/requirements.in # brax # chex # flax # jaxopt + # mujoco-mjx # optax # orbax-checkpoint -jaxlib==0.4.7+cuda11.cudnn82 +jax-cuda12-pjrt==0.4.28 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +jaxlib==0.4.28+cuda12.cudnn89 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # chex # jax # jaxopt + # mujoco-mjx # optax # orbax-checkpoint -jaxopt==0.7 +jaxopt==0.8.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # flask # torch -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -169,33 +179,43 @@ mdurl==0.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # markdown-it-py -ml-dtypes==0.1.0 +ml-collections==0.1.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +ml-dtypes==0.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax # jaxlib + # tensorstore mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -msgpack==1.0.5 +msgpack==1.0.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax # orbax-checkpoint -mujoco==2.3.5 +mujoco==3.1.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -nest-asyncio==1.5.6 +nest-asyncio==1.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # orbax-checkpoint -networkx==3.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.24.3 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax @@ -215,6 +235,56 @@ numpy==1.24.3 # tensorboardx # tensorstore # trimesh +nvidia-cublas-cu12==12.5.2.13 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 +nvidia-cuda-cupti-cu12==12.5.39 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cuda-nvcc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cuda-nvrtc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # nvidia-cudnn-cu12 +nvidia-cuda-runtime-cu12==12.5.39 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cudnn-cu12==8.9.7.29 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cufft-cu12==11.2.3.18 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cusolver-cu12==11.6.2.40 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cusparse-cu12==12.4.1.24 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax + # nvidia-cusolver-cu12 +nvidia-nccl-cu12==2.21.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax + # nvidia-cufft-cu12 + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -223,32 +293,42 @@ opt-einsum==3.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax -optax==0.1.5 +optax==0.2.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # flax -orbax-checkpoint==0.2.2 +orbax-checkpoint==0.5.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # brax # flax -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==9.5.0 +packaging==24.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # tensorboardx +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -protobuf==4.23.2 +protobuf==4.25.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # orbax-checkpoint # tensorboardx +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.15.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -264,68 +344,71 @@ pytinyrenderer==0.0.14 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax + # ml-collections # omegaconf # orbax-checkpoint reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -rich==13.3.5 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax # voir -scipy==1.10.1 +scipy==1.13.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # jax # jaxlib # jaxopt + # mujoco-mjx six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens - # flask-cors - # tensorboardx -sympy==1.12 + # ml-collections +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -tensorboardx==2.5 +tensorboardx==2.6.2.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -tensorstore==0.1.37 +tensorstore==0.1.60 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax # orbax-checkpoint -toolz==0.12.0 +toolz==0.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # chex -torch==2.0.1+rocm5.4.2 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/brax/requirements.in - # pytorch-triton-rocm -trimesh==3.9.35 +trimesh==4.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -typing-extensions==4.6.2 + # mujoco-mjx +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # chex + # etils # flax # orbax-checkpoint # reactivex @@ -334,17 +417,16 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.10 - # via -r benchmarks/brax/requirements.in -werkzeug==2.3.4 +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/brax/requirements.in +werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -zipp==3.15.0 +zipp==3.19.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # importlib-metadata - # importlib-resources - -# The following packages are considered to be unsafe in a requirements file: -# setuptools + # etils diff --git a/benchmarks/brax/requirements.xpu.txt b/benchmarks/brax/requirements.xpu.txt new file mode 100644 index 000000000..41b63f8a5 --- /dev/null +++ b/benchmarks/brax/requirements.xpu.txt @@ -0,0 +1,448 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/brax/requirements.xpu.txt .pin/tmp-constraints-xpu-brax.txt benchmarks/brax/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +blinker==1.8.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flask +brax==0.10.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/brax/requirements.in +chex==0.1.86 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # optax +click==8.1.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flask +cloudpickle==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # gym +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +contextlib2==21.6.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ml-collections +dm-env==1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +dm-tree==0.1.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # dm-env +etils[epath,epy]==1.9.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch + # triton +flask==3.0.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # flask-cors +flask-cors==4.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +flax==0.8.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # etils + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +glfw==2.7.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # mujoco +grpcio==1.64.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +gym==0.26.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +gym-notices==0.0.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # gym +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # etils +itsdangerous==2.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flask +jax[cuda12]==0.4.28 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # flask + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +ml-collections==0.1.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +ml-dtypes==0.4.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +msgpack==1.0.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +nest-asyncio==1.6.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # orbax-checkpoint +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # chex + # dm-env + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # opt-einsum + # optax + # orbax-checkpoint + # scipy + # tensorboardx + # tensorstore + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +opt-einsum==3.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax +optax==0.2.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # flax +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # tensorboardx +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +protobuf==4.25.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # orbax-checkpoint + # tensorboardx +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pyopengl==3.1.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # mujoco +pytinyrenderer==0.0.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flax + # ml-collections + # omegaconf + # orbax-checkpoint +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flax + # voir +scipy==1.13.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens + # ml-collections +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +tensorboardx==2.6.2.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +tensorstore==0.1.60 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flax + # orbax-checkpoint +toolz==0.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # chex +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/brax/requirements.in +trimesh==4.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # mujoco-mjx +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # chex + # etils + # flax + # orbax-checkpoint + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/brax/requirements.in +werkzeug==3.0.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flask +zipp==3.19.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # etils diff --git a/benchmarks/dlrm/requirements.cuda.txt b/benchmarks/dlrm/requirements.cuda.txt new file mode 100644 index 000000000..968d9359e --- /dev/null +++ b/benchmarks/dlrm/requirements.cuda.txt @@ -0,0 +1,359 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/dlrm/requirements.cuda.txt .pin/tmp-constraints-cuda-dlrm.txt benchmarks/dlrm/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +docker==7.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +docstring-parser==0.8.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +fbgemm-gpu==0.7.0+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchrec +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # torchx + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # torchx +future==1.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +graphviz==0.20.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchviz +grpcio==1.64.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +importlib-metadata==7.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +joblib==1.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # scikit-learn +lightning-utilities==0.11.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchmetrics +markdown==3.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +mypy-extensions==1.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # typing-inspect +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in + # fbgemm-gpu + # onnx + # scikit-learn + # scipy + # tensorboard + # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +onnx==1.16.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # lightning-utilities + # torchmetrics +protobuf==4.25.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # onnx + # tensorboard +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pydot==2.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyparsing==3.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pydot +pyre-extensions==0.0.30 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf + # torchx +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # docker +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +scikit-learn==1.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +scipy==1.13.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # scikit-learn +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens + # tensorboard +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tabulate==0.9.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +tensorboard==2.17.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +tensorboard-data-server==0.7.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +threadpoolctl==3.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # scikit-learn +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in + # torchmetrics + # torchviz +torchmetrics==1.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchrec +torchrec==0.7.0+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +torchviz==0.0.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +torchx==0.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in + # torchrec +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # lightning-utilities + # pyre-extensions + # reactivex + # torch + # typing-inspect +typing-inspect==0.9.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pyre-extensions +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # docker + # requests + # torchx +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/dlrm/requirements.in +werkzeug==3.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +zipp==3.19.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # importlib-metadata + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/dlrm/requirements.hpu.txt b/benchmarks/dlrm/requirements.hpu.txt index f13db0603..253c4bec6 100644 --- a/benchmarks/dlrm/requirements.hpu.txt +++ b/benchmarks/dlrm/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/dlrm/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-dlrm.txt milabench/benchmarks/dlrm/requirements.in +# pip-compile --output-file=benchmarks/dlrm/requirements.hpu.txt .pin/tmp-constraints-hpu-dlrm.txt benchmarks/dlrm/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + absl-py==2.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -16,7 +20,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -24,11 +28,11 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera -docker==7.0.0 +docker==7.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torchx @@ -56,7 +60,9 @@ fsspec==2024.3.1 # torch # torchx future==1.0.0 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in giving==0.4.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -66,7 +72,7 @@ graphviz==0.20.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torchviz -grpcio==1.63.0 +grpcio==1.64.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # tensorboard @@ -121,7 +127,8 @@ networkx==3.3 # torch numpy==1.26.4 # via - # -r milabench/benchmarks/dlrm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in # fbgemm-gpu # onnx # scikit-learn @@ -171,7 +178,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -184,19 +191,20 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -onnx==1.16.0 - # via -r milabench/benchmarks/dlrm/requirements.in +onnx==1.16.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt - # docker # lightning-utilities # torchmetrics -protobuf==5.26.1 +protobuf==4.25.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # onnx @@ -210,7 +218,9 @@ ptera==1.4.1 # -c .pin/../.pin/constraints-hpu-torch.txt # voir pydot==2.0.0 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in pygments==2.18.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -236,7 +246,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # docker @@ -244,9 +254,11 @@ rich==13.7.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -scikit-learn==1.4.2 - # via -r milabench/benchmarks/dlrm/requirements.in -scipy==1.13.0 +scikit-learn==1.5.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in +scipy==1.13.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # scikit-learn @@ -255,7 +267,7 @@ six==1.16.0 # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens # tensorboard -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch @@ -263,8 +275,10 @@ tabulate==0.9.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torchx -tensorboard==2.16.2 - # via -r milabench/benchmarks/dlrm/requirements.in +tensorboard==2.17.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in tensorboard-data-server==0.7.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -273,9 +287,10 @@ threadpoolctl==3.5.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # scikit-learn -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/dlrm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in # torchmetrics # torchviz torchmetrics==1.0.3 @@ -283,20 +298,27 @@ torchmetrics==1.0.3 # -c .pin/../.pin/constraints-hpu-torch.txt # torchrec torchrec==0.7.0 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in torchviz==0.0.2 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in torchx==0.5.0 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in tqdm==4.66.4 # via - # -r milabench/benchmarks/dlrm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in # torchrec -triton==2.3.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # lightning-utilities @@ -318,13 +340,16 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/dlrm/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/dlrm/requirements.in werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # tensorboard -zipp==3.18.1 +zipp==3.19.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # importlib-metadata diff --git a/benchmarks/dlrm/requirements.rocm.txt b/benchmarks/dlrm/requirements.rocm.txt index a36f32986..1095c63d1 100644 --- a/benchmarks/dlrm/requirements.rocm.txt +++ b/benchmarks/dlrm/requirements.rocm.txt @@ -2,11 +2,14 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/dlrm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-dlrm.txt benchmarks/dlrm/requirements.in +# pip-compile --output-file=benchmarks/dlrm/requirements.rocm.txt .pin/tmp-constraints-rocm-dlrm.txt benchmarks/dlrm/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -absl-py==2.0.0 +absl-py==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard @@ -18,11 +21,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -cachetools==5.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth -certifi==2023.7.22 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests @@ -30,15 +29,11 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera -docker==6.1.3 +docker==7.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx @@ -50,70 +45,59 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -fbgemm-gpu==0.5.0 +fbgemm-gpu==0.7.0+rocm6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchrec -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch # torchx -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch # torchx -future==0.18.3 - # via -r benchmarks/dlrm/requirements.in +future==1.0.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -google-auth==2.23.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth-oauthlib - # tensorboard -google-auth-oauthlib==1.1.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # tensorboard -graphviz==0.20.1 +graphviz==0.20.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchviz -grpcio==1.59.2 +grpcio==1.64.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -idna==3.4 +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -importlib-metadata==6.8.0 +importlib-metadata==7.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -joblib==1.3.2 +joblib==1.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn -lightning-utilities==0.9.0 +lightning-utilities==0.11.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchmetrics -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -markdown==3.5.1 +markdown==3.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard @@ -121,7 +105,7 @@ markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -138,12 +122,13 @@ mypy-extensions==1.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # typing-inspect -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/dlrm/requirements.in # fbgemm-gpu # onnx @@ -151,47 +136,41 @@ numpy==1.26.1 # scipy # tensorboard # torchmetrics -oauthlib==3.2.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests-oauthlib omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -onnx==1.15.0 - # via -r benchmarks/dlrm/requirements.in -ovld==0.3.2 +onnx==1.16.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # docker # lightning-utilities # torchmetrics -protobuf==4.23.4 +protobuf==4.25.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # onnx # tensorboard -ptera==1.4.1 +psutil==5.9.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pyasn1==0.5.0 +ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # pyasn1-modules - # rsa -pyasn1-modules==0.3.0 + # voir +pydot==2.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth -pydot==1.4.2 - # via -r benchmarks/dlrm/requirements.in -pygments==2.16.1 + # -r benchmarks/dlrm/requirements.in +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -199,7 +178,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pyparsing==3.1.1 +pyparsing==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pydot @@ -207,7 +186,7 @@ pyre-extensions==0.0.30 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -220,27 +199,19 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # docker - # requests-oauthlib - # tensorboard -requests-oauthlib==1.3.1 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth-oauthlib -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -rsa==4.9 +scikit-learn==1.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth -scikit-learn==1.3.2 - # via -r benchmarks/dlrm/requirements.in -scipy==1.11.3 + # -r benchmarks/dlrm/requirements.in +scipy==1.13.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn @@ -249,7 +220,7 @@ six==1.16.0 # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens # tensorboard -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -257,37 +228,46 @@ tabulate==0.9.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx -tensorboard==2.15.1 - # via -r benchmarks/dlrm/requirements.in +tensorboard==2.17.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in tensorboard-data-server==0.7.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -threadpoolctl==3.2.0 +threadpoolctl==3.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/dlrm/requirements.in - # pytorch-triton-rocm # torchmetrics # torchviz torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchrec -torchrec==0.5.0 - # via -r benchmarks/dlrm/requirements.in +torchrec==0.7.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in torchviz==0.0.2 - # via -r benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in torchx==0.5.0 - # via -r benchmarks/dlrm/requirements.in -tqdm==4.66.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/dlrm/requirements.in # torchrec -typing-extensions==4.8.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # lightning-utilities @@ -309,17 +289,16 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/dlrm/requirements.in -websocket-client==1.6.4 +voir==0.2.14 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # docker -werkzeug==3.0.1 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/dlrm/requirements.in +werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -zipp==3.17.0 +zipp==3.19.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # importlib-metadata diff --git a/benchmarks/dlrm/requirements.xpu.txt b/benchmarks/dlrm/requirements.xpu.txt index 1204a3a4e..6671afb1a 100644 --- a/benchmarks/dlrm/requirements.xpu.txt +++ b/benchmarks/dlrm/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/dlrm/requirements.xpu.txt .pin/tmp-constraints-xpu-dlrm.txt benchmarks/dlrm/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com absl-py==2.1.0 # via @@ -18,7 +21,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # requests @@ -26,11 +29,11 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera -docker==7.0.0 +docker==7.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchx @@ -42,16 +45,17 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -fbgemm-gpu==0.6.0 +fbgemm-gpu==0.7.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchrec -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch # torchx -fsspec==2024.2.0 + # triton +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -69,7 +73,7 @@ graphviz==0.20.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchviz -grpcio==1.62.2 +grpcio==1.64.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # tensorboard @@ -81,11 +85,11 @@ importlib-metadata==7.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchx -jinja2==3.1.3 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -joblib==1.4.0 +joblib==1.4.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # scikit-learn @@ -118,7 +122,7 @@ mypy-extensions==1.0.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # typing-inspect -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -132,11 +136,63 @@ numpy==1.26.4 # scipy # tensorboard # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -onnx==1.16.0 +onnx==1.16.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in @@ -144,17 +200,20 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # docker # lightning-utilities # torchmetrics -protobuf==5.26.1 +protobuf==4.25.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # onnx # tensorboard +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -163,7 +222,7 @@ pydot==2.0.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -188,7 +247,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # docker @@ -196,11 +255,11 @@ rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -scikit-learn==1.4.2 +scikit-learn==1.5.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in -scipy==1.13.0 +scipy==1.13.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # scikit-learn @@ -209,7 +268,7 @@ six==1.16.0 # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens # tensorboard -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -217,7 +276,7 @@ tabulate==0.9.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchx -tensorboard==2.16.2 +tensorboard==2.17.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in @@ -225,11 +284,11 @@ tensorboard-data-server==0.7.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # tensorboard -threadpoolctl==3.4.0 +threadpoolctl==3.5.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # scikit-learn -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt @@ -240,7 +299,7 @@ torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchrec -torchrec==0.6.0 +torchrec==0.7.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in @@ -252,12 +311,16 @@ torchx==0.5.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in -tqdm==4.66.2 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in # torchrec -typing-extensions==4.11.0 +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # lightning-utilities @@ -279,16 +342,16 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/dlrm/requirements.in -werkzeug==3.0.2 +werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # tensorboard -zipp==3.18.1 +zipp==3.19.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # importlib-metadata diff --git a/benchmarks/flops/requirements.cuda.txt b/benchmarks/flops/requirements.cuda.txt new file mode 100644 index 000000000..55ffc7fa8 --- /dev/null +++ b/benchmarks/flops/requirements.cuda.txt @@ -0,0 +1,210 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/flops/requirements.cuda.txt .pin/tmp-constraints-cuda-flops.txt benchmarks/flops/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchcompat +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/flops/requirements.in + # torchcompat + # torchvision +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/flops/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/flops/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/flops/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.hpu.txt b/benchmarks/flops/requirements.hpu.txt index 77d89923e..619bc2e5c 100644 --- a/benchmarks/flops/requirements.hpu.txt +++ b/benchmarks/flops/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/flops/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-flops.txt milabench/benchmarks/flops/requirements.in +# pip-compile --output-file=benchmarks/flops/requirements.hpu.txt .pin/tmp-constraints-hpu-flops.txt benchmarks/flops/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -109,7 +113,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -162,25 +166,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/flops/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/flops/requirements.in + # torchcompat # torchvision -torchcompat==0.0.1 - # via -r milabench/benchmarks/flops/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/flops/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/flops/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/flops/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/flops/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/flops/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -189,5 +202,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/flops/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.rocm.txt b/benchmarks/flops/requirements.rocm.txt index 23d10b701..f626e4d8d 100644 --- a/benchmarks/flops/requirements.rocm.txt +++ b/benchmarks/flops/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/flops/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-flops.txt benchmarks/flops/requirements.in +# pip-compile --output-file=benchmarks/flops/requirements.rocm.txt .pin/tmp-constraints-rocm-flops.txt benchmarks/flops/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,19 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,11 +63,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,19 +75,23 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +99,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +111,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,31 +119,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/flops/requirements.in - # pytorch-triton-rocm + # torchcompat # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/flops/requirements.in -tqdm==4.66.1 - # via -r benchmarks/flops/requirements.in -typing-extensions==4.8.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/flops/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/flops/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # -r benchmarks/flops/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/flops/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.xpu.txt b/benchmarks/flops/requirements.xpu.txt index 7308ad6fd..9b42646ea 100644 --- a/benchmarks/flops/requirements.xpu.txt +++ b/benchmarks/flops/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/flops/requirements.xpu.txt .pin/tmp-constraints-xpu-flops.txt benchmarks/flops/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera @@ -30,11 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -fsspec==2024.2.0 + # triton +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -43,11 +39,11 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -idna==3.7 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -jinja2==3.1.3 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -67,7 +63,7 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # sympy -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -75,6 +71,58 @@ numpy==1.26.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -87,11 +135,15 @@ pillow==10.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -107,10 +159,6 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # torchvision rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -119,39 +167,45 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/flops/requirements.in + # torchcompat # torchvision -torchvision==0.16.0a0+cxx11.abi +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/flops/requirements.in -tqdm==4.66.2 +torchvision==0.18.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt # -r benchmarks/flops/requirements.in -typing-extensions==4.11.0 +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/flops/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt diff --git a/benchmarks/huggingface/requirements.cuda.txt b/benchmarks/huggingface/requirements.cuda.txt new file mode 100644 index 000000000..962f943cb --- /dev/null +++ b/benchmarks/huggingface/requirements.cuda.txt @@ -0,0 +1,245 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/huggingface/requirements.cuda.txt .pin/tmp-constraints-cuda-hf.txt benchmarks/huggingface/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # torch + # transformers + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # transformers +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/huggingface/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/huggingface/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # reactivex + # torch +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.hpu.txt b/benchmarks/huggingface/requirements.hpu.txt index 042be0525..70511f523 100644 --- a/benchmarks/huggingface/requirements.hpu.txt +++ b/benchmarks/huggingface/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/huggingface/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-hf.txt milabench/benchmarks/huggingface/requirements.in +# pip-compile --output-file=benchmarks/huggingface/requirements.hpu.txt .pin/tmp-constraints-hpu-hf.txt benchmarks/huggingface/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -20,7 +24,7 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -45,7 +49,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-hpu-torch.txt # ptera # voir -huggingface-hub==0.23.0 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # tokenizers @@ -125,7 +129,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -142,7 +146,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -173,11 +177,11 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -regex==2024.4.28 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -194,7 +198,7 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch @@ -202,20 +206,24 @@ tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -torch==2.3.0 - # via -r milabench/benchmarks/huggingface/requirements.in +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/huggingface/requirements.in tqdm==4.66.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub # transformers -transformers==4.40.2 - # via -r milabench/benchmarks/huggingface/requirements.in -triton==2.3.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/huggingface/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -229,5 +237,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/huggingface/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.rocm.txt b/benchmarks/huggingface/requirements.rocm.txt index 4e39b0c45..983a5611a 100644 --- a/benchmarks/huggingface/requirements.rocm.txt +++ b/benchmarks/huggingface/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/huggingface/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-hf.txt benchmarks/huggingface/requirements.in +# pip-compile --output-file=benchmarks/huggingface/requirements.rocm.txt .pin/tmp-constraints-rocm-hf.txt benchmarks/huggingface/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,7 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests @@ -22,11 +25,7 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,14 +33,14 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pytorch-triton-rocm # torch # transformers -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -51,28 +50,24 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -huggingface-hub==0.17.3 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tokenizers # transformers -idna==3.4 +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -84,11 +79,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers @@ -96,20 +91,24 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # transformers +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -117,7 +116,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -131,20 +130,20 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -regex==2023.10.3 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # transformers -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -safetensors==0.4.0 +safetensors==0.4.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers @@ -152,26 +151,28 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -tokenizers==0.14.1 +tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/huggingface/requirements.in - # pytorch-triton-rocm -tqdm==4.66.1 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # transformers -transformers==4.35.0 - # via -r benchmarks/huggingface/requirements.in -typing-extensions==4.8.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/huggingface/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -185,5 +186,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/huggingface/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.xpu.txt b/benchmarks/huggingface/requirements.xpu.txt new file mode 100644 index 000000000..d1a96c409 --- /dev/null +++ b/benchmarks/huggingface/requirements.xpu.txt @@ -0,0 +1,246 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/huggingface/requirements.xpu.txt .pin/tmp-constraints-xpu-hf.txt benchmarks/huggingface/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # torch + # transformers + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # transformers +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/huggingface/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/huggingface/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # reactivex + # torch +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/llama/requirements.cuda.txt b/benchmarks/llama/requirements.cuda.txt new file mode 100644 index 000000000..e5d798adf --- /dev/null +++ b/benchmarks/llama/requirements.cuda.txt @@ -0,0 +1,344 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/llama/requirements.cuda.txt .pin/tmp-constraints-cuda-llm.txt benchmarks/llama/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +dill==0.3.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # multiprocess +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +fairscale==0.4.13 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # torch + # transformers + # triton +fire==0.6.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # fairscale + # pandas + # pyarrow + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # transformers +pandas==2.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyarrow==16.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +python-dateutil==2.9.0.post0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +pytz==2024.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +sentencepiece==0.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens + # fire + # python-dateutil +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +termcolor==2.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # fire +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in + # fairscale +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # reactivex + # torch +tzdata==2024.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/llama/requirements.in +xxhash==3.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp diff --git a/benchmarks/llama/requirements.hpu.txt b/benchmarks/llama/requirements.hpu.txt index 68275ff41..95b6d14b4 100644 --- a/benchmarks/llama/requirements.hpu.txt +++ b/benchmarks/llama/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/llama/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-llm.txt milabench/benchmarks/llama/requirements.in +# pip-compile --output-file=benchmarks/llama/requirements.hpu.txt .pin/tmp-constraints-hpu-llm.txt benchmarks/llama/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -21,15 +25,11 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -async-timeout==4.0.3 - # via - # -c .pin/../.pin/constraints-hpu-torch.txt - # aiohttp attrs==23.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # aiohttp -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -37,12 +37,14 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera -datasets==2.19.1 - # via -r milabench/benchmarks/llama/requirements.in +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in dill==0.3.8 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -53,7 +55,9 @@ executing==1.2.0 # -c .pin/../.pin/constraints-hpu-torch.txt # varname fairscale==0.4.13 - # via -r milabench/benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in filelock==3.14.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -63,7 +67,9 @@ filelock==3.14.0 # transformers # triton fire==0.6.0 - # via -r milabench/benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in frozenlist==1.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -80,7 +86,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-hpu-torch.txt # ptera # voir -huggingface-hub==0.23.0 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -175,7 +181,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -192,7 +198,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -210,7 +216,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -pyarrow==16.0.0 +pyarrow==16.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -245,11 +251,11 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -regex==2024.4.28 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -264,14 +270,16 @@ safetensors==0.4.3 # -c .pin/../.pin/constraints-hpu-torch.txt # transformers sentencepiece==0.2.0 - # via -r milabench/benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens # fire # python-dateutil -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch @@ -283,9 +291,10 @@ tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/llama/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in # fairscale tqdm==4.66.4 # via @@ -293,13 +302,15 @@ tqdm==4.66.4 # datasets # huggingface-hub # transformers -transformers==4.40.2 - # via -r milabench/benchmarks/llama/requirements.in -triton==2.3.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -317,8 +328,11 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/llama/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/llama/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/llama/requirements.rocm.txt b/benchmarks/llama/requirements.rocm.txt index eb26e2fa9..2339d207e 100644 --- a/benchmarks/llama/requirements.rocm.txt +++ b/benchmarks/llama/requirements.rocm.txt @@ -2,185 +2,291 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/llama/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-llm.txt benchmarks/llama/requirements.in +# pip-compile --output-file=benchmarks/llama/requirements.rocm.txt .pin/tmp-constraints-rocm-llm.txt benchmarks/llama/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -aiohttp==3.8.6 +aiohttp==3.9.5 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # fsspec aiosignal==1.3.1 - # via aiohttp + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # aiohttp antlr4-python3-runtime==4.9.3 - # via omegaconf + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # omegaconf asttokens==2.4.1 - # via giving -async-timeout==4.0.3 - # via aiohttp -attrs==23.1.0 - # via aiohttp -certifi==2023.7.22 - # via requests -charset-normalizer==3.3.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.27.7 - # via pytorch-triton-rocm -codefind==0.1.3 - # via ptera -datasets==2.14.6 - # via -r benchmarks/llama/requirements.in -dill==0.3.7 +charset-normalizer==3.3.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # ptera +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in +dill==0.3.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # multiprocess executing==1.2.0 - # via varname + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # varname fairscale==0.4.13 - # via -r benchmarks/llama/requirements.in -filelock==3.13.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets # huggingface-hub # pytorch-triton-rocm # torch # transformers -fire==0.5.0 - # via -r benchmarks/llama/requirements.in -frozenlist==1.4.0 +fire==0.6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.10.0 +fsspec[http]==2024.3.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub # torch giving==0.4.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -huggingface-hub==0.17.3 +huggingface-hub==0.23.3 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # tokenizers # transformers -idna==3.4 +idna==3.7 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # requests # yarl -jinja2==3.1.2 - # via torch -lit==17.0.4 - # via pytorch-triton-rocm +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch markdown-it-py==3.0.0 - # via rich -markupsafe==2.1.3 - # via jinja2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jinja2 mdurl==0.1.2 - # via markdown-it-py + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # markdown-it-py mpmath==1.3.0 - # via sympy -multidict==6.0.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # yarl -multiprocess==0.70.15 - # via datasets -networkx==3.2.1 - # via torch -numpy==1.26.1 +multiprocess==0.70.16 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +networkx==3.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # fairscale # pandas # pyarrow # transformers omegaconf==2.3.0 - # via voir -ovld==0.3.2 - # via voir -packaging==23.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub # transformers -pandas==2.1.2 - # via datasets +pandas==2.2.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 - # via voir -pyarrow==14.0.0 - # via datasets -pygments==2.16.1 - # via rich + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +pyarrow==16.1.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # rich pynvml==11.5.0 - # via voir -python-dateutil==2.8.2 - # via pandas -pytorch-triton-rocm==2.1.0 - # via torch -pytz==2023.3.post1 - # via pandas + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +python-dateutil==2.9.0.post0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pandas +pytorch-triton-rocm==2.3.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +pytz==2024.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pandas pyyaml==6.0.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub # omegaconf # transformers reactivex==4.0.4 - # via giving -regex==2023.10.3 - # via transformers -requests==2.31.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets - # fsspec # huggingface-hub # transformers -rich==13.6.0 - # via voir -safetensors==0.4.0 - # via transformers -sentencepiece==0.1.99 - # via -r benchmarks/llama/requirements.in +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # transformers +sentencepiece==0.2.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in six==1.16.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens # fire # python-dateutil -sympy==1.12 - # via torch -termcolor==2.3.0 - # via fire -tokenizers==0.14.1 - # via transformers -torch==2.1.0+rocm5.6 +sympy==1.12.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +termcolor==2.4.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # fire +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # transformers +torch==2.3.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/llama/requirements.in # fairscale - # pytorch-triton-rocm -tqdm==4.66.1 +tqdm==4.66.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub # transformers -transformers==4.35.0 - # via -r benchmarks/llama/requirements.in -typing-extensions==4.8.0 +transformers==4.41.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # reactivex # torch -tzdata==2023.3 - # via pandas -urllib3==2.0.7 - # via requests +tzdata==2024.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pandas +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # requests varname==0.10.0 - # via giving -voir==0.2.11 - # via -r benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/llama/requirements.in xxhash==3.4.1 - # via datasets -yarl==1.9.2 - # via aiohttp + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # aiohttp diff --git a/benchmarks/llama/requirements.xpu.txt b/benchmarks/llama/requirements.xpu.txt index b33512ae3..2009c62f2 100644 --- a/benchmarks/llama/requirements.xpu.txt +++ b/benchmarks/llama/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/llama/requirements.xpu.txt .pin/tmp-constraints-xpu-llm.txt benchmarks/llama/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com aiohttp==3.9.5 # via @@ -27,7 +30,7 @@ attrs==23.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # requests @@ -35,11 +38,11 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera -datasets==2.18.0 +datasets==2.19.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/llama/requirements.in @@ -56,13 +59,14 @@ fairscale==0.4.13 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/llama/requirements.in -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets # huggingface-hub # torch # transformers + # triton fire==0.6.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -72,7 +76,7 @@ frozenlist==1.4.1 # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp # aiosignal -fsspec[http]==2024.2.0 +fsspec[http]==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -83,7 +87,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -huggingface-hub==0.22.2 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -94,7 +98,7 @@ idna==3.7 # -c .pin/../.pin/constraints-xpu-torch.txt # requests # yarl -jinja2==3.1.3 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -123,7 +127,7 @@ multiprocess==0.70.16 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -135,6 +139,58 @@ numpy==1.26.4 # pandas # pyarrow # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -143,7 +199,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -153,11 +209,15 @@ pandas==2.2.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pyarrow==15.0.2 +pyarrow==16.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -165,7 +225,7 @@ pyarrow-hotfix==0.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -192,11 +252,11 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -regex==2024.4.16 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-xpu-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -220,7 +280,7 @@ six==1.16.0 # asttokens # fire # python-dateutil -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -232,23 +292,27 @@ tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # transformers -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/llama/requirements.in # fairscale -tqdm==4.66.2 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets # huggingface-hub # transformers -transformers==4.40.0 +transformers==4.41.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/llama/requirements.in -typing-extensions==4.11.0 +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # huggingface-hub @@ -266,7 +330,7 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt diff --git a/benchmarks/rwkv/requirements.cuda.txt b/benchmarks/rwkv/requirements.cuda.txt new file mode 100644 index 000000000..b48d463d5 --- /dev/null +++ b/benchmarks/rwkv/requirements.cuda.txt @@ -0,0 +1,280 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/rwkv/requirements.cuda.txt .pin/tmp-constraints-cuda-rwkv.txt benchmarks/rwkv/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pytorch-lightning + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +hjson==3.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +lightning-utilities==0.11.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pytorch-lightning + # torchmetrics +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # yarl +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +ninja==1.11.1.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed + # pytorch-lightning + # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # lightning-utilities + # pytorch-lightning + # torchmetrics +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +py-cpuinfo==9.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +pydantic==1.10.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # voir +pytorch-lightning==1.9.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf + # pytorch-lightning +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed + # pytorch-lightning + # torchmetrics +torchmetrics==1.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pytorch-lightning +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # pytorch-lightning +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # lightning-utilities + # pydantic + # pytorch-lightning + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/rwkv/requirements.in +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/rwkv/requirements.hpu.txt b/benchmarks/rwkv/requirements.hpu.txt index 200592d9b..57ef862cd 100644 --- a/benchmarks/rwkv/requirements.hpu.txt +++ b/benchmarks/rwkv/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/rwkv/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-rwkv.txt milabench/benchmarks/rwkv/requirements.in +# pip-compile --output-file=benchmarks/rwkv/requirements.hpu.txt .pin/tmp-constraints-hpu-rwkv.txt benchmarks/rwkv/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -20,20 +24,18 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -async-timeout==4.0.3 - # via - # -c .pin/../.pin/constraints-hpu-torch.txt - # aiohttp attrs==23.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # aiohttp -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera deepspeed==0.14.2 - # via -r milabench/benchmarks/rwkv/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -102,11 +104,13 @@ networkx==3.3 # torch ninja==1.11.1.1 # via - # -r milabench/benchmarks/rwkv/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed numpy==1.26.4 # via - # -r milabench/benchmarks/rwkv/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning # torchmetrics @@ -153,7 +157,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -170,7 +174,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed @@ -192,7 +196,8 @@ py-cpuinfo==9.0.0 # deepspeed pydantic==1.10.15 # via - # -r milabench/benchmarks/rwkv/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed pygments==2.18.0 # via @@ -204,7 +209,9 @@ pynvml==11.5.0 # deepspeed # voir pytorch-lightning==1.9.5 - # via -r milabench/benchmarks/rwkv/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -222,13 +229,14 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/rwkv/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning # torchmetrics @@ -241,11 +249,11 @@ tqdm==4.66.4 # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed # pytorch-lightning -triton==2.3.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # lightning-utilities @@ -257,8 +265,11 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/rwkv/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/rwkv/requirements.in yarl==1.9.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/rwkv/requirements.rocm.txt b/benchmarks/rwkv/requirements.rocm.txt index e97d63520..73b447ea0 100644 --- a/benchmarks/rwkv/requirements.rocm.txt +++ b/benchmarks/rwkv/requirements.rocm.txt @@ -2,11 +2,14 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/rwkv/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-rwkv.txt benchmarks/rwkv/requirements.in +# pip-compile --output-file=benchmarks/rwkv/requirements.rocm.txt .pin/tmp-constraints-rocm-rwkv.txt benchmarks/rwkv/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -aiohttp==3.8.6 +aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # fsspec @@ -22,48 +25,33 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -async-timeout==4.0.3 +attrs==23.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp -attrs==23.1.0 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # aiohttp -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # aiohttp - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 + # ptera +deepspeed==0.14.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # ptera -deepspeed==0.12.2 - # via -r benchmarks/rwkv/requirements.in + # -r benchmarks/rwkv/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -frozenlist==1.4.0 +frozenlist==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.10.0 +fsspec[http]==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning @@ -77,29 +65,24 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -idna==3.4 +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests # yarl -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lightning-utilities==0.9.0 +lightning-utilities==0.11.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning # torchmetrics -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -111,21 +94,23 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -multidict==6.0.4 +multidict==6.0.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # yarl -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch ninja==1.11.1.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/rwkv/requirements.in # deepspeed -numpy==1.26.1 +numpy==1.26.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning @@ -134,21 +119,22 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed # lightning-utilities # pytorch-lightning # torchmetrics -psutil==5.9.6 +psutil==5.9.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -157,11 +143,12 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pydantic==1.10.13 +pydantic==1.10.15 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/rwkv/requirements.in # deepspeed -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -171,8 +158,10 @@ pynvml==11.5.0 # deepspeed # voir pytorch-lightning==1.9.5 - # via -r benchmarks/rwkv/requirements.in -pytorch-triton-rocm==2.1.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/rwkv/requirements.in +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -185,11 +174,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # fsspec -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -197,27 +182,27 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning - # pytorch-triton-rocm # torchmetrics torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning -tqdm==4.66.1 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed # pytorch-lightning -typing-extensions==4.8.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # lightning-utilities @@ -225,17 +210,19 @@ typing-extensions==4.8.0 # pytorch-lightning # reactivex # torch -urllib3==1.26.18 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/rwkv/requirements.in -yarl==1.9.2 +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/rwkv/requirements.in +yarl==1.9.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/rwkv/requirements.xpu.txt b/benchmarks/rwkv/requirements.xpu.txt new file mode 100644 index 000000000..0c9fd4e6f --- /dev/null +++ b/benchmarks/rwkv/requirements.xpu.txt @@ -0,0 +1,281 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/rwkv/requirements.xpu.txt .pin/tmp-constraints-xpu-rwkv.txt benchmarks/rwkv/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch + # triton +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pytorch-lightning + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +hjson==3.1.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +idna==3.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +lightning-utilities==0.11.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pytorch-lightning + # torchmetrics +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + # yarl +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +ninja==1.11.1.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed + # pytorch-lightning + # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # lightning-utilities + # pytorch-lightning + # torchmetrics +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +py-cpuinfo==9.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +pydantic==1.10.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # voir +pytorch-lightning==1.9.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf + # pytorch-lightning +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed + # pytorch-lightning + # torchmetrics +torchmetrics==1.0.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pytorch-lightning +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # pytorch-lightning +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # lightning-utilities + # pydantic + # pytorch-lightning + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/rwkv/requirements.in +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/stargan/prepare.py b/benchmarks/stargan/prepare.py old mode 100644 new mode 100755 diff --git a/benchmarks/stargan/requirements.cuda.txt b/benchmarks/stargan/requirements.cuda.txt index f4d74306e..be5e6570d 100644 --- a/benchmarks/stargan/requirements.cuda.txt +++ b/benchmarks/stargan/requirements.cuda.txt @@ -6,6 +6,7 @@ # --extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html --trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 @@ -29,7 +30,7 @@ filelock==3.14.0 # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -fsspec==2024.6.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch diff --git a/benchmarks/stargan/requirements.hpu.txt b/benchmarks/stargan/requirements.hpu.txt index a787f68c5..bd6008ae7 100644 --- a/benchmarks/stargan/requirements.hpu.txt +++ b/benchmarks/stargan/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/stargan/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-stargan.txt milabench/benchmarks/stargan/requirements.in +# pip-compile --output-file=benchmarks/stargan/requirements.hpu.txt .pin/tmp-constraints-hpu-stargan.txt benchmarks/stargan/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -60,7 +64,8 @@ networkx==3.3 # torch numpy==1.26.4 # via - # -r milabench/benchmarks/stargan/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/stargan/requirements.in # torchvision nvidia-cublas-cu12==12.1.3.1 # via @@ -105,7 +110,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -158,21 +163,24 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/stargan/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/stargan/requirements.in # torchvision -torchvision==0.18.0 - # via -r milabench/benchmarks/stargan/requirements.in -triton==2.3.0 +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/stargan/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -181,5 +189,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/stargan/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/stargan/requirements.in diff --git a/benchmarks/stargan/requirements.rocm.txt b/benchmarks/stargan/requirements.rocm.txt index d2b904c55..0ad8b961a 100644 --- a/benchmarks/stargan/requirements.rocm.txt +++ b/benchmarks/stargan/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/stargan/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-stargan.txt benchmarks/stargan/requirements.in +# pip-compile --output-file=benchmarks/stargan/requirements.rocm.txt .pin/tmp-constraints-rocm-stargan.txt benchmarks/stargan/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,15 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,31 +59,36 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/stargan/requirements.in # torchvision omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +96,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +108,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,29 +116,30 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/stargan/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/stargan/requirements.in -typing-extensions==4.8.0 +torchvision==0.18.1+rocm6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -r benchmarks/stargan/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/stargan/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/stargan/requirements.in diff --git a/benchmarks/stargan/requirements.xpu.txt b/benchmarks/stargan/requirements.xpu.txt index e1e1ac4f5..7c74b1879 100644 --- a/benchmarks/stargan/requirements.xpu.txt +++ b/benchmarks/stargan/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/stargan/requirements.xpu.txt .pin/tmp-constraints-xpu-stargan.txt benchmarks/stargan/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera @@ -30,11 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -fsspec==2024.2.0 + # triton +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -43,11 +39,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -idna==3.7 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -jinja2==3.1.3 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -67,7 +59,7 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # sympy -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -76,6 +68,58 @@ numpy==1.26.4 # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/stargan/requirements.in # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -88,11 +132,15 @@ pillow==10.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -108,10 +156,6 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # torchvision rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -120,35 +164,35 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/stargan/requirements.in # torchvision -torchvision==0.16.0a0+cxx11.abi +torchvision==0.18.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/stargan/requirements.in -typing-extensions==4.11.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt diff --git a/benchmarks/super-slomo/requirements.cuda.txt b/benchmarks/super-slomo/requirements.cuda.txt new file mode 100644 index 000000000..0f646fb30 --- /dev/null +++ b/benchmarks/super-slomo/requirements.cuda.txt @@ -0,0 +1,206 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/super-slomo/requirements.cuda.txt .pin/tmp-constraints-cuda-super-slomo.txt benchmarks/super-slomo/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in + # opencv-python + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +opencv-python==4.10.0.82 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in + # torchvision +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.hpu.txt b/benchmarks/super-slomo/requirements.hpu.txt index 0169803e3..e4e71df05 100644 --- a/benchmarks/super-slomo/requirements.hpu.txt +++ b/benchmarks/super-slomo/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/super-slomo/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-super-slomo.txt milabench/benchmarks/super-slomo/requirements.in +# pip-compile --output-file=benchmarks/super-slomo/requirements.hpu.txt .pin/tmp-constraints-hpu-super-slomo.txt benchmarks/super-slomo/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -60,7 +64,8 @@ networkx==3.3 # torch numpy==1.26.4 # via - # -r milabench/benchmarks/super-slomo/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in # opencv-python # torchvision nvidia-cublas-cu12==12.1.3.1 @@ -106,7 +111,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -119,8 +124,10 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -opencv-python==4.9.0.80 - # via -r milabench/benchmarks/super-slomo/requirements.in +opencv-python==4.10.0.82 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -161,23 +168,28 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/super-slomo/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in # torchvision -torchvision==0.18.0 - # via -r milabench/benchmarks/super-slomo/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/super-slomo/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -186,5 +198,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/super-slomo/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.rocm.txt b/benchmarks/super-slomo/requirements.rocm.txt index 02ff9f070..426ca5898 100644 --- a/benchmarks/super-slomo/requirements.rocm.txt +++ b/benchmarks/super-slomo/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/super-slomo/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-super-slomo.txt benchmarks/super-slomo/requirements.in +# pip-compile --output-file=benchmarks/super-slomo/requirements.rocm.txt .pin/tmp-constraints-rocm-super-slomo.txt benchmarks/super-slomo/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,15 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,12 +59,13 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/super-slomo/requirements.in # opencv-python # torchvision @@ -89,21 +73,27 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -opencv-python==4.8.1.78 - # via -r benchmarks/super-slomo/requirements.in -ovld==0.3.2 +opencv-python==4.10.0.82 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/super-slomo/requirements.in +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -111,7 +101,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -123,11 +113,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -135,31 +121,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/super-slomo/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/super-slomo/requirements.in -tqdm==4.66.1 - # via -r benchmarks/super-slomo/requirements.in -typing-extensions==4.8.0 +torchvision==0.18.1+rocm6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -r benchmarks/super-slomo/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # -r benchmarks/super-slomo/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/super-slomo/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.xpu.txt b/benchmarks/super-slomo/requirements.xpu.txt index afce02611..0c0db1f7b 100644 --- a/benchmarks/super-slomo/requirements.xpu.txt +++ b/benchmarks/super-slomo/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/super-slomo/requirements.xpu.txt .pin/tmp-constraints-xpu-super-slomo.txt benchmarks/super-slomo/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera @@ -30,11 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -fsspec==2024.2.0 + # triton +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -43,11 +39,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -idna==3.7 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -jinja2==3.1.3 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -67,7 +59,7 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # sympy -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -77,11 +69,63 @@ numpy==1.26.4 # -r benchmarks/super-slomo/requirements.in # opencv-python # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -opencv-python==4.9.0.80 +opencv-python==4.10.0.82 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/super-slomo/requirements.in @@ -93,11 +137,15 @@ pillow==10.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -113,10 +161,6 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # torchvision rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -125,39 +169,39 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/super-slomo/requirements.in # torchvision -torchvision==0.16.0a0+cxx11.abi +torchvision==0.18.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/super-slomo/requirements.in -tqdm==4.66.2 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/super-slomo/requirements.in -typing-extensions==4.11.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt diff --git a/benchmarks/timm/requirements.cuda.txt b/benchmarks/timm/requirements.cuda.txt new file mode 100644 index 000000000..0ae4b86f3 --- /dev/null +++ b/benchmarks/timm/requirements.cuda.txt @@ -0,0 +1,237 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/timm/requirements.cuda.txt .pin/tmp-constraints-cuda-timm.txt benchmarks/timm/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in + # huggingface-hub + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in + # torchvision +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # reactivex + # torch +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.hpu.txt b/benchmarks/timm/requirements.hpu.txt index 3c3e53a5a..0e857f02b 100644 --- a/benchmarks/timm/requirements.hpu.txt +++ b/benchmarks/timm/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/timm/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-timm.txt milabench/benchmarks/timm/requirements.in +# pip-compile --output-file=benchmarks/timm/requirements.hpu.txt .pin/tmp-constraints-hpu-timm.txt benchmarks/timm/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -20,7 +24,7 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -44,8 +48,10 @@ giving==0.4.2 # -c .pin/../.pin/constraints-hpu-torch.txt # ptera # voir -huggingface-hub==0.23.0 - # via -r milabench/benchmarks/timm/requirements.in +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in idna==3.7 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -121,7 +127,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -138,7 +144,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -164,14 +170,15 @@ pynvml==11.5.0 # voir pyyaml==6.0.1 # via - # -r milabench/benchmarks/timm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in # huggingface-hub # omegaconf reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -180,30 +187,35 @@ rich==13.7.1 # -c .pin/../.pin/constraints-hpu-torch.txt # voir safetensors==0.4.3 - # via -r milabench/benchmarks/timm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/timm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in # torchvision -torchvision==0.18.0 - # via -r milabench/benchmarks/timm/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in tqdm==4.66.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub -triton==2.3.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -217,5 +229,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/timm/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.rocm.txt b/benchmarks/timm/requirements.rocm.txt index 6b15125f5..c94ddb3cf 100644 --- a/benchmarks/timm/requirements.rocm.txt +++ b/benchmarks/timm/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/timm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-timm.txt benchmarks/timm/requirements.in +# pip-compile --output-file=benchmarks/timm/requirements.rocm.txt .pin/tmp-constraints-rocm-timm.txt benchmarks/timm/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,7 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests @@ -22,11 +25,7 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,13 +33,13 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -50,25 +49,23 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -huggingface-hub==0.17.3 - # via -r benchmarks/timm/requirements.in -idna==3.4 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # -r benchmarks/timm/requirements.in +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # torch -lit==17.0.4 + # requests +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm + # torch markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -80,11 +77,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -92,23 +89,27 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -116,12 +117,13 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch pyyaml==6.0.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/timm/requirements.in # huggingface-hub # omegaconf @@ -129,37 +131,40 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -safetensors==0.4.0 - # via -r benchmarks/timm/requirements.in +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/timm/requirements.in six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/timm/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/timm/requirements.in -tqdm==4.66.1 +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/timm/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub -typing-extensions==4.8.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -173,5 +178,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/timm/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.xpu.txt b/benchmarks/timm/requirements.xpu.txt new file mode 100644 index 000000000..198a41877 --- /dev/null +++ b/benchmarks/timm/requirements.xpu.txt @@ -0,0 +1,239 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/timm/requirements.xpu.txt .pin/tmp-constraints-xpu-timm.txt benchmarks/timm/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/timm/requirements.in +idna==3.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/timm/requirements.in + # huggingface-hub + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/timm/requirements.in +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/timm/requirements.in + # torchvision +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/timm/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # reactivex + # torch +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/torchvision/requirements.cuda.txt b/benchmarks/torchvision/requirements.cuda.txt new file mode 100644 index 000000000..829d01147 --- /dev/null +++ b/benchmarks/torchvision/requirements.cuda.txt @@ -0,0 +1,210 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/torchvision/requirements.cuda.txt .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchcompat +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision/requirements.in + # torchcompat + # torchvision +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.hpu.txt b/benchmarks/torchvision/requirements.hpu.txt index 575ec65f6..e62dfdb37 100644 --- a/benchmarks/torchvision/requirements.hpu.txt +++ b/benchmarks/torchvision/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/torchvision/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-torchvision.txt milabench/benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision/requirements.hpu.txt .pin/tmp-constraints-hpu-torchvision.txt benchmarks/torchvision/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -109,7 +113,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -162,25 +166,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/torchvision/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in + # torchcompat # torchvision -torchcompat==0.0.1 - # via -r milabench/benchmarks/torchvision/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/torchvision/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/torchvision/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -189,5 +202,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/torchvision/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.rocm.txt b/benchmarks/torchvision/requirements.rocm.txt index 618dff8f8..a9f1128cf 100644 --- a/benchmarks/torchvision/requirements.rocm.txt +++ b/benchmarks/torchvision/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision/requirements.rocm.txt .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,19 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,11 +63,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,19 +75,23 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +99,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +111,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,31 +119,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/torchvision/requirements.in - # pytorch-triton-rocm + # torchcompat # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/torchvision/requirements.in -tqdm==4.66.1 - # via -r benchmarks/torchvision/requirements.in -typing-extensions==4.8.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # -r benchmarks/torchvision/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/torchvision/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.xpu.txt b/benchmarks/torchvision/requirements.xpu.txt index af75c7076..329ae4314 100644 --- a/benchmarks/torchvision/requirements.xpu.txt +++ b/benchmarks/torchvision/requirements.xpu.txt @@ -6,122 +6,207 @@ # --extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html --trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 - # via omegaconf + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf asttokens==2.4.1 - # via giving + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving codefind==0.1.6 - # via ptera + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera executing==1.2.0 - # via varname + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname filelock==3.14.0 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # torch # triton -fsspec==2024.6.0 - # via torch +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch giving==0.4.2 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir importlib-resources==6.4.0 - # via torchcompat + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchcompat jinja2==3.1.4 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch markdown-it-py==3.0.0 - # via rich + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich markupsafe==2.1.5 - # via jinja2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 mdurl==0.1.2 - # via markdown-it-py + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py mpmath==1.3.0 - # via sympy + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy networkx==3.3 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch numpy==1.26.4 - # via torchvision + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision nvidia-cublas-cu12==12.1.3.1 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # nvidia-cudnn-cu12 # nvidia-cusolver-cu12 # torch nvidia-cuda-cupti-cu12==12.1.105 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cuda-nvrtc-cu12==12.1.105 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cuda-runtime-cu12==12.1.105 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cudnn-cu12==8.9.2.26 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cufft-cu12==11.0.2.54 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-curand-cu12==10.3.2.106 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cusolver-cu12==11.4.5.107 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cusparse-cu12==12.1.0.106 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # nvidia-cusolver-cu12 # torch nvidia-nccl-cu12==2.20.5 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-nvjitlink-cu12==12.5.40 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # nvidia-cusolver-cu12 # nvidia-cusparse-cu12 nvidia-nvtx-cu12==12.1.105 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ovld==0.3.5 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir pillow==10.3.0 - # via torchvision + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision psutil==5.9.8 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir pygments==2.18.0 - # via rich + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich pynvml==11.5.0 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir pyyaml==6.0.1 - # via omegaconf + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf reactivex==4.0.4 - # via giving + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving rich==13.7.1 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir six==1.16.0 - # via asttokens + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens sympy==1.12.1 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch torch==2.3.1 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in # torchcompat # torchvision torchcompat==1.0.2 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in torchvision==0.18.1 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in tqdm==4.66.4 - # via -r benchmarks/torchvision/requirements.in + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/torchvision/requirements.in triton==2.3.1 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch typing-extensions==4.12.2 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # reactivex # torch varname==0.10.0 - # via giving + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving voir==0.2.14 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.cuda.txt b/benchmarks/torchvision_ddp/requirements.cuda.txt new file mode 100644 index 000000000..f9afea09e --- /dev/null +++ b/benchmarks/torchvision_ddp/requirements.cuda.txt @@ -0,0 +1,210 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.cuda.txt .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision_ddp/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchcompat +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat + # torchvision +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.hpu.txt b/benchmarks/torchvision_ddp/requirements.hpu.txt index 575ec65f6..87b9bc008 100644 --- a/benchmarks/torchvision_ddp/requirements.hpu.txt +++ b/benchmarks/torchvision_ddp/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/torchvision/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-torchvision.txt milabench/benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.hpu.txt .pin/tmp-constraints-hpu-torchvision.txt benchmarks/torchvision_ddp/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -109,7 +113,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -162,25 +166,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/torchvision/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat # torchvision -torchcompat==0.0.1 - # via -r milabench/benchmarks/torchvision/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/torchvision/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/torchvision/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -189,5 +202,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/torchvision/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.rocm.txt b/benchmarks/torchvision_ddp/requirements.rocm.txt index 618dff8f8..e3aac9376 100644 --- a/benchmarks/torchvision_ddp/requirements.rocm.txt +++ b/benchmarks/torchvision_ddp/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.rocm.txt .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision_ddp/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,19 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,11 +63,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,19 +75,23 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +99,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +111,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,31 +119,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via - # -r benchmarks/torchvision/requirements.in - # pytorch-triton-rocm + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/torchvision/requirements.in -tqdm==4.66.1 - # via -r benchmarks/torchvision/requirements.in -typing-extensions==4.8.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/torchvision/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.xpu.txt b/benchmarks/torchvision_ddp/requirements.xpu.txt new file mode 100644 index 000000000..acf8d589f --- /dev/null +++ b/benchmarks/torchvision_ddp/requirements.xpu.txt @@ -0,0 +1,212 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.xpu.txt .pin/tmp-constraints-xpu-torchvision.txt benchmarks/torchvision_ddp/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchcompat +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat + # torchvision +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/config/base.yaml b/config/base.yaml index dc138db22..917bdc116 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -202,7 +202,7 @@ resnet50: argv: --model: resnet50 - --batch-size: 64 + --batch-size: 256 --num-workers: "{cpu_per_gpu}" --loader: pytorch @@ -217,8 +217,7 @@ resnet50-noio: argv: --model: resnet50 - --batch-size: 64 - --num-workers: 8 + --batch-size: 256 --loader: synthetic_fixed resnet152-ddp: diff --git a/config/standard.yaml b/config/standard.yaml index c984b1571..3f43a8055 100644 --- a/config/standard.yaml +++ b/config/standard.yaml @@ -126,7 +126,7 @@ dlrm: weight: 1.0 rwkv: - enabled: true + enabled: false weight: 1.0 fp16: diff --git a/milabench/_version.py b/milabench/_version.py index 0f3a43042..57c79d91e 100644 --- a/milabench/_version.py +++ b/milabench/_version.py @@ -1,5 +1,5 @@ """This file is generated, do not modify""" -__tag__ = "v0.0.6-132-gf372aae" -__commit__ = "f372aae6e8e2d6fc2ec9dd738b3a3f0939f27d17" -__date__ = "2024-06-06 16:55:27 -0400" +__tag__ = "v0.0.6-140-g57343f1" +__commit__ = "57343f10ef2b4ce598011ee308ebd06b4c654495" +__date__ = "2024-06-10 11:52:37 -0400" diff --git a/milabench/multi.py b/milabench/multi.py index 1e9089e83..4ea76c701 100644 --- a/milabench/multi.py +++ b/milabench/multi.py @@ -243,6 +243,9 @@ async def do_pin( working_dir=here.parent ) + if not ig_constraint_path.exists(): + raise RuntimeError("Could not generate main requirements") + # Use master requirements to constrain the rest new_constraints = [ig_constraint_path, *constraints] for pack in packs: diff --git a/scripts/article/run_cuda.sh b/scripts/article/run_cuda.sh index 34c85b1d5..682e4a34e 100644 --- a/scripts/article/run_cuda.sh +++ b/scripts/article/run_cuda.sh @@ -15,7 +15,9 @@ install_prepare() { mkdir -p $MILABENCH_WORDIR cd $MILABENCH_WORDIR - virtualenv $MILABENCH_WORDIR/env + if [ ! -d "$MILABENCH_WORDIR/env" ]; then + virtualenv $MILABENCH_WORDIR/env + fi if [ ! -d "$MILABENCH_WORDIR/milabench" ]; then git clone https://github.com/mila-iqia/milabench.git -b intel @@ -52,7 +54,7 @@ install_prepare() { module load cuda/12.3.2 -if [ ! -d "$MILABENCH_WORDIR" ]; then +if [ ! -d "$MILABENCH_WORDIR/results" ]; then install_prepare else echo "Reusing previous install" diff --git a/scripts/config.yaml b/scripts/config.yaml deleted file mode 100644 index e69de29bb..000000000 diff --git a/scripts/multinode.sh b/scripts/multinode.sh deleted file mode 100644 index e69de29bb..000000000 diff --git a/scripts/update_pins.sh b/scripts/update_pins.sh index ad7a30f7b..c16f5db6a 100644 --- a/scripts/update_pins.sh +++ b/scripts/update_pins.sh @@ -1,7 +1,5 @@ - - MILABENCH_GPU_ARCH=cuda milabench pin -c constraints/cuda.txt --config config/standard.yaml --from-scratch MILABENCH_GPU_ARCH=rocm milabench pin -c constraints/rocm.txt --config config/standard.yaml --from-scratch MILABENCH_GPU_ARCH=xpu milabench pin -c constraints/xpu.txt --config config/standard.yaml --from-scratch -MILABENCH_GPU_ARCH=hpu milabench pin -c constraints/hpu.txt --config config/standard.yaml --from-scratch \ No newline at end of file +MILABENCH_GPU_ARCH=hpu milabench pin -c constraints/hpu.txt --config config/standard.yaml --from-scratch