diff --git a/.github/workflows/tests.yml b/.github/workflows/tests.yml index 8e0cb45d2..5eb69d068 100644 --- a/.github/workflows/tests.yml +++ b/.github/workflows/tests.yml @@ -76,6 +76,12 @@ jobs: poetry lock --no-update poetry install + - name: pin + run: | + MILABENCH_GPU_ARCH=cuda poetry run milabench pin -c constraints/cuda.txt --config config/standard.yaml + MILABENCH_GPU_ARCH=rocm poetry run milabench pin -c constraints/rocm.txt --config config/standard.yaml + git diff --stat + - name: tests run: | export PATH="/opt/rocm/bin:$PATH" diff --git a/.gitignore b/.gitignore index 6ff12f7c5..8e6de4a30 100644 --- a/.gitignore +++ b/.gitignore @@ -19,3 +19,7 @@ sqlite.db .pin-constraints-* .mongo/ .pin/tmp-* + + +.no_report +trash/ diff --git a/.pin/constraints-cuda-torch.txt b/.pin/constraints-cuda-torch.txt index 9db708f95..9662ad4d3 100644 --- a/.pin/constraints-cuda-torch.txt +++ b/.pin/constraints-cuda-torch.txt @@ -1,16 +1,16 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=.pin/constraints-cuda-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/dlrm/requirements.in benchmarks/huggingface/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in +# pip-compile --config=pyproject.toml --output-file=.pin/constraints-cuda-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/dlrm/requirements.in benchmarks/huggingface/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 -absl-py==1.4.0 +absl-py==2.0.0 # via tensorboard -accelerate==0.19.0 +accelerate==0.23.0 # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.4 +aiohttp==3.8.6 # via # datasets # fsspec @@ -18,62 +18,63 @@ aiosignal==1.3.1 # via aiohttp antlr4-python3-runtime==4.9.3 # via omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via giving -async-timeout==4.0.2 +async-timeout==4.0.3 # via aiohttp attrs==23.1.0 # via aiohttp cachetools==5.3.1 # via google-auth -certifi==2023.5.7 +certifi==2023.7.22 # via requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # aiohttp # requests -cmake==3.26.3 - # via triton codefind==0.1.3 # via ptera -datasets==2.12.0 +datasets==2.14.6 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.8.3 - # via -r benchmarks/accelerate_opt/requirements.in -dill==0.3.6 +deepspeed==0.11.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/rwkv/requirements.in +dill==0.3.7 # via # datasets # evaluate # multiprocess -docker==6.1.2 +docker==6.1.3 # via torchx -docstring-parser==0.8.1 +docstring-parser==0.15 # via torchx -evaluate==0.4.0 +evaluate==0.4.1 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via varname -fbgemm-gpu==0.4.1 +fbgemm-gpu==0.5.0+cu118 # via torchrec -filelock==3.12.0 +filelock==3.12.4 # via # huggingface-hub # torch # torchx # transformers # triton -frozenlist==1.3.3 +frozenlist==1.4.0 # via # aiohttp # aiosignal -fsspec[http]==2023.5.0 +fsspec[http]==2023.1.0 # via # datasets # evaluate # huggingface-hub # pytorch-lightning + # torch # torchx future==0.18.3 # via -r benchmarks/dlrm/requirements.in @@ -81,47 +82,45 @@ giving==0.4.2 # via # ptera # voir -google-auth==2.19.0 +google-auth==2.23.3 # via # google-auth-oauthlib # tensorboard -google-auth-oauthlib==1.0.0 +google-auth-oauthlib==1.1.0 # via tensorboard graphviz==0.20.1 # via torchviz -grpcio==1.54.2 +grpcio==1.59.0 # via tensorboard hjson==3.1.0 # via deepspeed -huggingface-hub==0.14.1 +huggingface-hub==0.17.3 # via # -r benchmarks/timm/requirements.in + # accelerate # datasets # evaluate + # tokenizers # transformers idna==3.4 # via # requests # yarl -importlib-metadata==6.6.0 - # via - # markdown - # torchx -iopath==0.1.10 - # via torchrec +importlib-metadata==6.8.0 + # via torchx jinja2==3.1.2 # via torch -joblib==1.2.0 +joblib==1.3.2 # via scikit-learn -lightning-utilities==0.8.0 - # via pytorch-lightning -lit==16.0.5 - # via triton -markdown==3.4.3 +lightning-utilities==0.9.0 + # via + # pytorch-lightning + # torchmetrics +markdown==3.5 # via tensorboard -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # jinja2 # werkzeug @@ -133,21 +132,21 @@ multidict==6.0.4 # via # aiohttp # yarl -multiprocess==0.70.14 +multiprocess==0.70.15 # via # datasets # evaluate mypy-extensions==1.0.0 # via typing-inspect -networkx==3.1 +networkx==3.2 # via torch -ninja==1.11.1 +ninja==1.11.1.1 # via # -r benchmarks/rwkv/requirements.in # deepspeed -numpy==1.24.3 +numpy==1.26.1 # via - # -r benchmarks/dlrm/requirements.in + # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # accelerate # datasets @@ -168,13 +167,13 @@ oauthlib==3.2.2 # via requests-oauthlib omegaconf==2.3.0 # via voir -onnx==1.14.0 +onnx==1.14.1 # via -r benchmarks/dlrm/requirements.in -opencv-python==4.7.0.72 +opencv-python==4.8.1.78 # via -r benchmarks/super-slomo/requirements.in ovld==0.3.2 # via voir -packaging==23.1 +packaging==23.2 # via # accelerate # datasets @@ -186,20 +185,17 @@ packaging==23.1 # pytorch-lightning # torchmetrics # transformers -pandas==2.0.2 +pandas==2.1.1 # via # datasets # evaluate - # torchrec -pillow==9.5.0 +pillow==10.1.0 # via torchvision -portalocker==2.7.0 - # via iopath -protobuf==4.23.2 +protobuf==4.23.4 # via # onnx # tensorboard -psutil==5.9.5 +psutil==5.9.6 # via # accelerate # deepspeed @@ -207,7 +203,7 @@ ptera==1.4.1 # via voir py-cpuinfo==9.0.0 # via deepspeed -pyarrow==12.0.0 +pyarrow==13.0.0 # via datasets pyasn1==0.5.0 # via @@ -215,15 +211,17 @@ pyasn1==0.5.0 # rsa pyasn1-modules==0.3.0 # via google-auth -pydantic==1.10.8 - # via deepspeed +pydantic==1.10.13 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed pydot==1.4.2 # via -r benchmarks/dlrm/requirements.in -pygments==2.15.1 +pygments==2.16.1 # via rich pynvml==11.5.0 # via voir -pyparsing==3.0.9 +pyparsing==3.1.1 # via pydot pyre-extensions==0.0.30 # via torchx @@ -231,9 +229,9 @@ python-dateutil==2.8.2 # via pandas pytorch-lightning==1.9.5 # via -r benchmarks/rwkv/requirements.in -pytz==2023.3 +pytz==2023.3.post1 # via pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -r benchmarks/timm/requirements.in # accelerate @@ -245,7 +243,7 @@ pyyaml==6.0 # transformers reactivex==4.0.4 # via giving -regex==2023.5.5 +regex==2023.10.3 # via transformers requests==2.31.0 # via @@ -262,44 +260,42 @@ requests==2.31.0 requests-oauthlib==1.3.1 # via google-auth-oauthlib responses==0.18.0 - # via - # datasets - # evaluate -rich==13.3.5 + # via evaluate +rich==13.6.0 # via # -r benchmarks/accelerate_opt/requirements.in # voir rsa==4.9 # via google-auth -safetensors==0.3.1 - # via -r benchmarks/timm/requirements.in -scikit-learn==1.2.2 +safetensors==0.4.0 + # via + # -r benchmarks/timm/requirements.in + # transformers +scikit-learn==1.3.2 # via -r benchmarks/dlrm/requirements.in -scipy==1.10.1 +scipy==1.11.3 # via scikit-learn six==1.16.0 # via # asttokens - # google-auth # python-dateutil + # tensorboard sympy==1.12 # via torch tabulate==0.9.0 - # via - # torchrec - # torchx -tensorboard==2.13.0 + # via torchx +tensorboard==2.15.0 # via -r benchmarks/dlrm/requirements.in -tensorboard-data-server==0.7.0 +tensorboard-data-server==0.7.1 # via tensorboard -threadpoolctl==3.1.0 +threadpoolctl==3.2.0 # via scikit-learn -tokenizers==0.13.3 +tokenizers==0.14.1 # via transformers -torch==2.0.1+cu118 +torch==2.1.0+cu118 # via - # -r benchmarks/dlrm/requirements.in - # -r benchmarks/timm/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/super-slomo/requirements.in # accelerate # deepspeed # pytorch-lightning @@ -307,45 +303,42 @@ torch==2.0.1+cu118 # torchmetrics # torchvision # torchviz - # triton -torchaudio==2.0.2+cu118 +torchaudio==2.1.0+cu118 # via -r benchmarks/accelerate_opt/requirements.in -torchmetrics==0.11.4 +torchmetrics==1.0.3 # via # pytorch-lightning # torchrec -torchrec==0.4.0 +torchrec==0.5.0+cu118 # via -r benchmarks/dlrm/requirements.in -torchvision==0.15.2+cu118 +torchvision==0.16.0+cu118 # via # -r benchmarks/super-slomo/requirements.in # -r benchmarks/torchvision/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in -torchx==0.5.0 +torchx==0.6.0 # via -r benchmarks/dlrm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via - # -r benchmarks/dlrm/requirements.in + # -r benchmarks/super-slomo/requirements.in # -r benchmarks/torchvision/requirements.in # datasets # deepspeed # evaluate # huggingface-hub - # iopath # pytorch-lightning # torchrec # transformers -transformers==4.29.2 +transformers==4.34.1 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/huggingface/requirements.in -triton==2.0.0 +triton==2.1.0 # via torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # huggingface-hub - # iopath # lightning-utilities # onnx # pydantic @@ -358,10 +351,9 @@ typing-inspect==0.9.0 # via pyre-extensions tzdata==2023.3 # via pandas -urllib3==1.26.16 +urllib3==1.26.18 # via # docker - # google-auth # requests # responses # torchx @@ -369,21 +361,19 @@ varname==0.10.0 # via giving voir==0.2.10 # via - # -r benchmarks/dlrm/requirements.in - # -r benchmarks/timm/requirements.in -websocket-client==1.5.2 + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/super-slomo/requirements.in +websocket-client==1.6.4 # via docker -werkzeug==2.3.4 - # via tensorboard -wheel==0.40.0 +werkzeug==3.0.0 # via tensorboard -xxhash==3.2.0 +xxhash==3.4.1 # via # datasets # evaluate yarl==1.9.2 # via aiohttp -zipp==3.15.0 +zipp==3.17.0 # via importlib-metadata # The following packages are considered to be unsafe in a requirements file: diff --git a/.pin/constraints-rocm-torch.txt b/.pin/constraints-rocm-torch.txt index efd674cce..486032fcf 100644 --- a/.pin/constraints-rocm-torch.txt +++ b/.pin/constraints-rocm-torch.txt @@ -1,16 +1,16 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=.pin/constraints-rocm-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/dlrm/requirements.in benchmarks/huggingface/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in +# pip-compile --config=pyproject.toml --output-file=.pin/constraints-rocm-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/dlrm/requirements.in benchmarks/huggingface/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ -absl-py==1.4.0 +absl-py==2.0.0 # via tensorboard -accelerate==0.19.0 +accelerate==0.23.0 # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.4 +aiohttp==3.8.6 # via # datasets # fsspec @@ -18,62 +18,65 @@ aiosignal==1.3.1 # via aiohttp antlr4-python3-runtime==4.9.3 # via omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via giving -async-timeout==4.0.2 +async-timeout==4.0.3 # via aiohttp attrs==23.1.0 # via aiohttp cachetools==5.3.1 # via google-auth -certifi==2023.5.7 +certifi==2023.7.22 # via requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # aiohttp # requests -cmake==3.26.3 +cmake==3.27.7 # via pytorch-triton-rocm codefind==0.1.3 # via ptera -datasets==2.12.0 +datasets==2.14.6 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.8.3 - # via -r benchmarks/accelerate_opt/requirements.in -dill==0.3.6 +deepspeed==0.11.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/rwkv/requirements.in +dill==0.3.7 # via # datasets # evaluate # multiprocess -docker==6.1.2 +docker==6.1.3 # via torchx -docstring-parser==0.8.1 +docstring-parser==0.15 # via torchx -evaluate==0.4.0 +evaluate==0.4.1 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via varname -fbgemm-gpu==0.4.1 +fbgemm-gpu==0.5.0 # via torchrec -filelock==3.12.0 +filelock==3.12.4 # via # huggingface-hub # pytorch-triton-rocm # torch # torchx # transformers -frozenlist==1.3.3 +frozenlist==1.4.0 # via # aiohttp # aiosignal -fsspec[http]==2023.5.0 +fsspec[http]==2023.1.0 # via # datasets # evaluate # huggingface-hub # pytorch-lightning + # torch # torchx future==0.18.3 # via -r benchmarks/dlrm/requirements.in @@ -81,47 +84,47 @@ giving==0.4.2 # via # ptera # voir -google-auth==2.19.0 +google-auth==2.23.3 # via # google-auth-oauthlib # tensorboard -google-auth-oauthlib==1.0.0 +google-auth-oauthlib==1.1.0 # via tensorboard graphviz==0.20.1 # via torchviz -grpcio==1.54.2 +grpcio==1.59.0 # via tensorboard hjson==3.1.0 # via deepspeed -huggingface-hub==0.14.1 +huggingface-hub==0.17.3 # via # -r benchmarks/timm/requirements.in + # accelerate # datasets # evaluate + # tokenizers # transformers idna==3.4 # via # requests # yarl -importlib-metadata==6.6.0 - # via - # markdown - # torchx -iopath==0.1.10 - # via torchrec +importlib-metadata==6.8.0 + # via torchx jinja2==3.1.2 # via torch -joblib==1.2.0 +joblib==1.3.2 # via scikit-learn -lightning-utilities==0.8.0 - # via pytorch-lightning -lit==16.0.5 +lightning-utilities==0.9.0 + # via + # pytorch-lightning + # torchmetrics +lit==17.0.3 # via pytorch-triton-rocm -markdown==3.4.3 +markdown==3.5 # via tensorboard -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # jinja2 # werkzeug @@ -133,26 +136,27 @@ multidict==6.0.4 # via # aiohttp # yarl -multiprocess==0.70.14 +multiprocess==0.70.15 # via # datasets # evaluate mypy-extensions==1.0.0 # via typing-inspect -networkx==3.1 +networkx==3.2 # via torch -ninja==1.11.1 +ninja==1.11.1.1 # via # -r benchmarks/rwkv/requirements.in # deepspeed -numpy==1.24.3 +numpy==1.26.1 # via - # -r benchmarks/dlrm/requirements.in + # -r benchmarks/stargan/requirements.in # -r benchmarks/super-slomo/requirements.in # accelerate # datasets # deepspeed # evaluate + # fbgemm-gpu # onnx # opencv-python # pandas @@ -168,13 +172,13 @@ oauthlib==3.2.2 # via requests-oauthlib omegaconf==2.3.0 # via voir -onnx==1.14.0 +onnx==1.14.1 # via -r benchmarks/dlrm/requirements.in -opencv-python==4.7.0.72 +opencv-python==4.8.1.78 # via -r benchmarks/super-slomo/requirements.in ovld==0.3.2 # via voir -packaging==23.1 +packaging==23.2 # via # accelerate # datasets @@ -186,20 +190,17 @@ packaging==23.1 # pytorch-lightning # torchmetrics # transformers -pandas==2.0.2 +pandas==2.1.1 # via # datasets # evaluate - # torchrec -pillow==9.5.0 +pillow==10.1.0 # via torchvision -portalocker==2.7.0 - # via iopath -protobuf==4.23.2 +protobuf==4.23.4 # via # onnx # tensorboard -psutil==5.9.5 +psutil==5.9.6 # via # accelerate # deepspeed @@ -207,7 +208,7 @@ ptera==1.4.1 # via voir py-cpuinfo==9.0.0 # via deepspeed -pyarrow==12.0.0 +pyarrow==13.0.0 # via datasets pyasn1==0.5.0 # via @@ -215,15 +216,17 @@ pyasn1==0.5.0 # rsa pyasn1-modules==0.3.0 # via google-auth -pydantic==1.10.8 - # via deepspeed +pydantic==1.10.13 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed pydot==1.4.2 # via -r benchmarks/dlrm/requirements.in -pygments==2.15.1 +pygments==2.16.1 # via rich pynvml==11.5.0 # via voir -pyparsing==3.0.9 +pyparsing==3.1.1 # via pydot pyre-extensions==0.0.30 # via torchx @@ -231,11 +234,11 @@ python-dateutil==2.8.2 # via pandas pytorch-lightning==1.9.5 # via -r benchmarks/rwkv/requirements.in -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via torch -pytz==2023.3 +pytz==2023.3.post1 # via pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -r benchmarks/timm/requirements.in # accelerate @@ -247,7 +250,7 @@ pyyaml==6.0 # transformers reactivex==4.0.4 # via giving -regex==2023.5.5 +regex==2023.10.3 # via transformers requests==2.31.0 # via @@ -264,43 +267,41 @@ requests==2.31.0 requests-oauthlib==1.3.1 # via google-auth-oauthlib responses==0.18.0 - # via - # datasets - # evaluate -rich==13.3.5 + # via evaluate +rich==13.6.0 # via # -r benchmarks/accelerate_opt/requirements.in # voir rsa==4.9 # via google-auth -safetensors==0.3.1 - # via -r benchmarks/timm/requirements.in -scikit-learn==1.2.2 +safetensors==0.4.0 + # via + # -r benchmarks/timm/requirements.in + # transformers +scikit-learn==1.3.2 # via -r benchmarks/dlrm/requirements.in -scipy==1.10.1 +scipy==1.11.3 # via scikit-learn six==1.16.0 # via # asttokens - # google-auth # python-dateutil + # tensorboard sympy==1.12 # via torch tabulate==0.9.0 - # via - # torchrec - # torchx -tensorboard==2.13.0 + # via torchx +tensorboard==2.15.0 # via -r benchmarks/dlrm/requirements.in -tensorboard-data-server==0.7.0 +tensorboard-data-server==0.7.1 # via tensorboard -threadpoolctl==3.1.0 +threadpoolctl==3.2.0 # via scikit-learn -tokenizers==0.13.3 +tokenizers==0.14.1 # via transformers -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via - # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/stargan/requirements.in # -r benchmarks/torchvision/requirements.in # accelerate # deepspeed @@ -310,23 +311,23 @@ torch==2.0.1+rocm5.4.2 # torchmetrics # torchvision # torchviz -torchaudio==2.0.2+rocm5.4.2 +torchaudio==2.1.0+rocm5.6 # via -r benchmarks/accelerate_opt/requirements.in -torchmetrics==0.11.4 +torchmetrics==1.0.3 # via # pytorch-lightning # torchrec -torchrec==0.4.0 +torchrec==0.5.0 # via -r benchmarks/dlrm/requirements.in -torchvision==0.15.2+rocm5.4.2 +torchvision==0.16.0+rocm5.6 # via - # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/stargan/requirements.in # -r benchmarks/torchvision/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in -torchx==0.5.0 +torchx==0.6.0 # via -r benchmarks/dlrm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -r benchmarks/dlrm/requirements.in # -r benchmarks/torchvision/requirements.in @@ -334,18 +335,16 @@ tqdm==4.65.0 # deepspeed # evaluate # huggingface-hub - # iopath # pytorch-lightning # torchrec # transformers -transformers==4.29.2 +transformers==4.34.1 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/huggingface/requirements.in -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # huggingface-hub - # iopath # lightning-utilities # onnx # pydantic @@ -358,10 +357,9 @@ typing-inspect==0.9.0 # via pyre-extensions tzdata==2023.3 # via pandas -urllib3==1.26.16 +urllib3==1.26.18 # via # docker - # google-auth # requests # responses # torchx @@ -369,21 +367,19 @@ varname==0.10.0 # via giving voir==0.2.10 # via - # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/stargan/requirements.in # -r benchmarks/torchvision/requirements.in -websocket-client==1.5.2 +websocket-client==1.6.4 # via docker -werkzeug==2.3.4 - # via tensorboard -wheel==0.40.0 +werkzeug==3.0.0 # via tensorboard -xxhash==3.2.0 +xxhash==3.4.1 # via # datasets # evaluate yarl==1.9.2 # via aiohttp -zipp==3.15.0 +zipp==3.17.0 # via importlib-metadata # The following packages are considered to be unsafe in a requirements file: diff --git a/benchmarks/accelerate_opt/requirements.cuda.txt b/benchmarks/accelerate_opt/requirements.cuda.txt index 1f4194c6b..0886bde19 100644 --- a/benchmarks/accelerate_opt/requirements.cuda.txt +++ b/benchmarks/accelerate_opt/requirements.cuda.txt @@ -1,14 +1,14 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/accelerate_opt/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-opt.txt benchmarks/accelerate_opt/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/accelerate_opt/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-opt.txt benchmarks/accelerate_opt/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 -accelerate==0.19.0 +accelerate==0.23.0 # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.4 +aiohttp==3.8.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -21,11 +21,11 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -async-timeout==4.0.2 +async-timeout==4.0.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp @@ -33,59 +33,56 @@ attrs==23.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # requests -cmake==3.26.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton codefind==0.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera -datasets==2.12.0 +datasets==2.14.6 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.8.3 +deepspeed==0.11.1 # via -r benchmarks/accelerate_opt/requirements.in -dill==0.3.6 +dill==0.3.7 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets # evaluate # multiprocess -evaluate==0.4.0 +evaluate==0.4.1 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # torch # transformers # triton -frozenlist==1.3.3 +frozenlist==1.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.5.0 +fsspec[http]==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets # evaluate # huggingface-hub + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -95,11 +92,13 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed -huggingface-hub==0.14.1 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate # datasets # evaluate + # tokenizers # transformers idna==3.4 # via @@ -110,15 +109,11 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -135,20 +130,20 @@ multidict==6.0.4 # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # yarl -multiprocess==0.70.14 +multiprocess==0.70.15 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets # evaluate -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -ninja==1.11.1 +ninja==1.11.1.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed -numpy==1.24.3 +numpy==1.26.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # accelerate @@ -167,7 +162,7 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # accelerate @@ -176,16 +171,16 @@ packaging==23.1 # evaluate # huggingface-hub # transformers -pandas==2.0.2 +pandas==2.1.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets # evaluate -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision -psutil==5.9.5 +psutil==5.9.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # accelerate @@ -198,15 +193,15 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed -pyarrow==12.0.0 +pyarrow==13.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets -pydantic==1.10.8 +pydantic==1.10.13 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -218,11 +213,11 @@ python-dateutil==2.8.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # pandas -pytz==2023.3 +pytz==2023.3.post1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # accelerate @@ -234,7 +229,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -regex==2023.5.5 +regex==2023.10.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # transformers @@ -251,12 +246,15 @@ requests==2.31.0 responses==0.18.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # datasets # evaluate -rich==13.3.5 +rich==13.6.0 # via # -r benchmarks/accelerate_opt/requirements.in # voir +safetensors==0.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers six==1.16.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -266,23 +264,22 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -tokenizers==0.13.3 +tokenizers==0.14.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # transformers -torch==2.0.1+cu118 +torch==2.1.0+cu118 # via # -r benchmarks/accelerate_opt/requirements.in # accelerate # deepspeed # torchaudio # torchvision - # triton -torchaudio==2.0.2+cu118 +torchaudio==2.1.0+cu118 # via -r benchmarks/accelerate_opt/requirements.in -torchvision==0.15.2+cu118 +torchvision==0.16.0+cu118 # via -r benchmarks/accelerate_opt/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets @@ -290,13 +287,13 @@ tqdm==4.65.0 # evaluate # huggingface-hub # transformers -transformers==4.29.2 +transformers==4.34.1 # via -r benchmarks/accelerate_opt/requirements.in -triton==2.0.0 +triton==2.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub @@ -307,7 +304,7 @@ tzdata==2023.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # pandas -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests @@ -318,7 +315,7 @@ varname==0.10.0 # giving voir==0.2.10 # via -r benchmarks/accelerate_opt/requirements.in -xxhash==3.2.0 +xxhash==3.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # datasets diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index 09cbce64f..64f34b7b1 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -1,14 +1,14 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/accelerate_opt/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-opt.txt benchmarks/accelerate_opt/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/accelerate_opt/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-opt.txt benchmarks/accelerate_opt/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ -accelerate==0.19.0 +accelerate==0.23.0 # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.4 +aiohttp==3.8.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets @@ -21,11 +21,11 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -async-timeout==4.0.2 +async-timeout==4.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp @@ -33,16 +33,16 @@ attrs==23.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # requests -cmake==3.26.3 +cmake==3.27.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm @@ -50,42 +50,43 @@ codefind==0.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera -datasets==2.12.0 +datasets==2.14.6 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.8.3 +deepspeed==0.11.1 # via -r benchmarks/accelerate_opt/requirements.in -dill==0.3.6 +dill==0.3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate # multiprocess -evaluate==0.4.0 +evaluate==0.4.1 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pytorch-triton-rocm # torch # transformers -frozenlist==1.3.3 +frozenlist==1.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.5.0 +fsspec[http]==2023.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate # huggingface-hub + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -95,11 +96,13 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -huggingface-hub==0.14.1 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # accelerate # datasets # evaluate + # tokenizers # transformers idna==3.4 # via @@ -110,15 +113,15 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==16.0.5 +lit==17.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -135,20 +138,20 @@ multidict==6.0.4 # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # yarl -multiprocess==0.70.14 +multiprocess==0.70.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -ninja==1.11.1 +ninja==1.11.1.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -numpy==1.24.3 +numpy==1.26.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -167,7 +170,7 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -176,16 +179,16 @@ packaging==23.1 # evaluate # huggingface-hub # transformers -pandas==2.0.2 +pandas==2.1.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision -psutil==5.9.5 +psutil==5.9.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -198,15 +201,15 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pyarrow==12.0.0 +pyarrow==13.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets -pydantic==1.10.8 +pydantic==1.10.13 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -218,15 +221,15 @@ python-dateutil==2.8.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pytz==2023.3 +pytz==2023.3.post1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -238,7 +241,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -regex==2023.5.5 +regex==2023.10.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers @@ -255,12 +258,15 @@ requests==2.31.0 responses==0.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # datasets # evaluate -rich==13.3.5 +rich==13.6.0 # via # -r benchmarks/accelerate_opt/requirements.in # voir +safetensors==0.4.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # transformers six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -270,11 +276,11 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -tokenizers==0.13.3 +tokenizers==0.14.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via # -r benchmarks/accelerate_opt/requirements.in # accelerate @@ -282,11 +288,11 @@ torch==2.0.1+rocm5.4.2 # pytorch-triton-rocm # torchaudio # torchvision -torchaudio==2.0.2+rocm5.4.2 +torchaudio==2.1.0+rocm5.6 # via -r benchmarks/accelerate_opt/requirements.in -torchvision==0.15.2+rocm5.4.2 +torchvision==0.16.0+rocm5.6 # via -r benchmarks/accelerate_opt/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets @@ -294,9 +300,9 @@ tqdm==4.65.0 # evaluate # huggingface-hub # transformers -transformers==4.29.2 +transformers==4.34.1 # via -r benchmarks/accelerate_opt/requirements.in -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -307,7 +313,7 @@ tzdata==2023.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests @@ -318,7 +324,7 @@ varname==0.10.0 # giving voir==0.2.10 # via -r benchmarks/accelerate_opt/requirements.in -xxhash==3.2.0 +xxhash==3.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets diff --git a/benchmarks/dlrm/requirements.cuda.txt b/benchmarks/dlrm/requirements.cuda.txt index 98abcb0cf..def37f9e5 100644 --- a/benchmarks/dlrm/requirements.cuda.txt +++ b/benchmarks/dlrm/requirements.cuda.txt @@ -1,12 +1,12 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/dlrm/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-dlrm.txt benchmarks/dlrm/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/dlrm/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-dlrm.txt benchmarks/dlrm/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 -absl-py==1.4.0 +absl-py==2.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # tensorboard @@ -14,7 +14,7 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving @@ -22,27 +22,23 @@ cachetools==5.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # google-auth -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -cmake==3.26.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton codefind==0.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera -docker==6.1.2 +docker==6.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchx -docstring-parser==0.8.1 +docstring-parser==0.15 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchx @@ -50,19 +46,20 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -fbgemm-gpu==0.4.1 +fbgemm-gpu==0.5.0+cu118 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchrec -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch # torchx # triton -fsspec==2023.5.0 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt + # torch # torchx future==0.18.3 # via -r benchmarks/dlrm/requirements.in @@ -71,12 +68,12 @@ giving==0.4.2 # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -google-auth==2.19.0 +google-auth==2.23.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # google-auth-oauthlib # tensorboard -google-auth-oauthlib==1.0.0 +google-auth-oauthlib==1.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # tensorboard @@ -84,7 +81,7 @@ graphviz==0.20.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchviz -grpcio==1.54.2 +grpcio==1.59.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # tensorboard @@ -92,36 +89,31 @@ idna==3.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -importlib-metadata==6.6.0 +importlib-metadata==6.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # markdown # torchx -iopath==0.1.10 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # torchrec jinja2==3.1.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -joblib==1.2.0 +joblib==1.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # scikit-learn -lit==16.0.5 +lightning-utilities==0.9.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # triton -markdown==3.4.3 + # torchmetrics +markdown==3.5 # via # -c .pin/../.pin/constraints-cuda-torch.txt # tensorboard -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -138,15 +130,14 @@ mypy-extensions==1.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # typing-inspect -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -r benchmarks/dlrm/requirements.in # onnx - # pandas # scikit-learn # scipy # tensorboard @@ -159,26 +150,19 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -onnx==1.14.0 +onnx==1.14.1 # via -r benchmarks/dlrm/requirements.in ovld==0.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # docker + # lightning-utilities # torchmetrics -pandas==2.0.2 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # torchrec -portalocker==2.7.0 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # iopath -protobuf==4.23.2 +protobuf==4.23.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # onnx @@ -198,7 +182,7 @@ pyasn1-modules==0.3.0 # google-auth pydot==1.4.2 # via -r benchmarks/dlrm/requirements.in -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -206,7 +190,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pyparsing==3.0.9 +pyparsing==3.1.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # pydot @@ -214,15 +198,7 @@ pyre-extensions==0.0.30 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchx -python-dateutil==2.8.2 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # pandas -pytz==2023.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf @@ -241,7 +217,7 @@ requests-oauthlib==1.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # google-auth-oauthlib -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir @@ -249,9 +225,9 @@ rsa==4.9 # via # -c .pin/../.pin/constraints-cuda-torch.txt # google-auth -scikit-learn==1.2.2 +scikit-learn==1.3.2 # via -r benchmarks/dlrm/requirements.in -scipy==1.10.1 +scipy==1.11.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # scikit-learn @@ -259,8 +235,7 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # asttokens - # google-auth - # python-dateutil + # tensorboard sympy==1.12 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -268,47 +243,44 @@ sympy==1.12 tabulate==0.9.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # torchrec # torchx -tensorboard==2.13.0 +tensorboard==2.15.0 # via -r benchmarks/dlrm/requirements.in -tensorboard-data-server==0.7.0 +tensorboard-data-server==0.7.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # tensorboard -threadpoolctl==3.1.0 +threadpoolctl==3.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # scikit-learn -torch==2.0.1+cu118 +torch==2.1.0+cu118 # via # -r benchmarks/dlrm/requirements.in # torchmetrics # torchviz - # triton -torchmetrics==0.11.4 +torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchrec -torchrec==0.4.0 +torchrec==0.5.0+cu118 # via -r benchmarks/dlrm/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in -torchx==0.5.0 +torchx==0.6.0 # via -r benchmarks/dlrm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -r benchmarks/dlrm/requirements.in - # iopath # torchrec -triton==2.0.0 +triton==2.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt - # iopath + # lightning-utilities # onnx # pyre-extensions # reactivex @@ -318,15 +290,10 @@ typing-inspect==0.9.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # pyre-extensions -tzdata==2023.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # pandas -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-cuda-torch.txt # docker - # google-auth # requests # torchx varname==0.10.0 @@ -335,19 +302,15 @@ varname==0.10.0 # giving voir==0.2.10 # via -r benchmarks/dlrm/requirements.in -websocket-client==1.5.2 +websocket-client==1.6.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # docker -werkzeug==2.3.4 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # tensorboard -wheel==0.40.0 +werkzeug==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # tensorboard -zipp==3.15.0 +zipp==3.17.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # importlib-metadata diff --git a/benchmarks/dlrm/requirements.rocm.txt b/benchmarks/dlrm/requirements.rocm.txt index a8d29307e..1249e82b2 100644 --- a/benchmarks/dlrm/requirements.rocm.txt +++ b/benchmarks/dlrm/requirements.rocm.txt @@ -1,12 +1,12 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/dlrm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-dlrm.txt benchmarks/dlrm/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/dlrm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-dlrm.txt benchmarks/dlrm/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ -absl-py==1.4.0 +absl-py==2.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard @@ -14,7 +14,7 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving @@ -22,15 +22,15 @@ cachetools==5.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # google-auth -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.26.3 +cmake==3.27.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm @@ -38,11 +38,11 @@ codefind==0.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera -docker==6.1.2 +docker==6.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx -docstring-parser==0.8.1 +docstring-parser==0.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx @@ -50,19 +50,20 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -fbgemm-gpu==0.4.1 +fbgemm-gpu==0.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchrec -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch # torchx -fsspec==2023.5.0 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # torch # torchx future==0.18.3 # via -r benchmarks/dlrm/requirements.in @@ -71,12 +72,12 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -google-auth==2.19.0 +google-auth==2.23.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # google-auth-oauthlib # tensorboard -google-auth-oauthlib==1.0.0 +google-auth-oauthlib==1.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard @@ -84,7 +85,7 @@ graphviz==0.20.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchviz -grpcio==1.54.2 +grpcio==1.59.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard @@ -92,36 +93,35 @@ idna==3.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -importlib-metadata==6.6.0 +importlib-metadata==6.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # markdown # torchx -iopath==0.1.10 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchrec jinja2==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -joblib==1.2.0 +joblib==1.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn -lit==16.0.5 +lightning-utilities==0.9.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torchmetrics +lit==17.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm -markdown==3.4.3 +markdown==3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -138,15 +138,15 @@ mypy-extensions==1.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # typing-inspect -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -r benchmarks/dlrm/requirements.in + # fbgemm-gpu # onnx - # pandas # scikit-learn # scipy # tensorboard @@ -159,26 +159,19 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -onnx==1.14.0 +onnx==1.14.1 # via -r benchmarks/dlrm/requirements.in ovld==0.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # docker + # lightning-utilities # torchmetrics -pandas==2.0.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchrec -portalocker==2.7.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # iopath -protobuf==4.23.2 +protobuf==4.23.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # onnx @@ -198,7 +191,7 @@ pyasn1-modules==0.3.0 # google-auth pydot==1.4.2 # via -r benchmarks/dlrm/requirements.in -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -206,7 +199,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pyparsing==3.0.9 +pyparsing==3.1.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pydot @@ -214,19 +207,11 @@ pyre-extensions==0.0.30 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx -python-dateutil==2.8.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pandas -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pytz==2023.3 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf @@ -245,7 +230,7 @@ requests-oauthlib==1.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # google-auth-oauthlib -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -253,9 +238,9 @@ rsa==4.9 # via # -c .pin/../.pin/constraints-rocm-torch.txt # google-auth -scikit-learn==1.2.2 +scikit-learn==1.3.2 # via -r benchmarks/dlrm/requirements.in -scipy==1.10.1 +scipy==1.11.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn @@ -263,8 +248,7 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens - # google-auth - # python-dateutil + # tensorboard sympy==1.12 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -272,43 +256,41 @@ sympy==1.12 tabulate==0.9.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # torchrec # torchx -tensorboard==2.13.0 +tensorboard==2.15.0 # via -r benchmarks/dlrm/requirements.in -tensorboard-data-server==0.7.0 +tensorboard-data-server==0.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -threadpoolctl==3.1.0 +threadpoolctl==3.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via # -r benchmarks/dlrm/requirements.in # pytorch-triton-rocm # torchmetrics # torchviz -torchmetrics==0.11.4 +torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchrec -torchrec==0.4.0 +torchrec==0.5.0 # via -r benchmarks/dlrm/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in -torchx==0.5.0 +torchx==0.6.0 # via -r benchmarks/dlrm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -r benchmarks/dlrm/requirements.in - # iopath # torchrec -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # iopath + # lightning-utilities # onnx # pyre-extensions # reactivex @@ -318,15 +300,10 @@ typing-inspect==0.9.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pyre-extensions -tzdata==2023.3 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pandas -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # docker - # google-auth # requests # torchx varname==0.10.0 @@ -335,19 +312,15 @@ varname==0.10.0 # giving voir==0.2.10 # via -r benchmarks/dlrm/requirements.in -websocket-client==1.5.2 +websocket-client==1.6.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # docker -werkzeug==2.3.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # tensorboard -wheel==0.40.0 +werkzeug==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -zipp==3.15.0 +zipp==3.17.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # importlib-metadata diff --git a/benchmarks/huggingface/requirements.cuda.txt b/benchmarks/huggingface/requirements.cuda.txt index 70053a636..1e2890b37 100644 --- a/benchmarks/huggingface/requirements.cuda.txt +++ b/benchmarks/huggingface/requirements.cuda.txt @@ -1,8 +1,8 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/huggingface/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-hf.txt benchmarks/huggingface/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/huggingface/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-hf.txt benchmarks/huggingface/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -cmake==3.26.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton codefind==0.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -34,25 +30,27 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # torch # transformers # triton -fsspec==2023.5.0 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -huggingface-hub==0.14.1 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt + # tokenizers # transformers idna==3.4 # via @@ -62,15 +60,11 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -82,11 +76,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # transformers @@ -98,7 +92,7 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub @@ -107,7 +101,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -115,7 +109,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub @@ -125,7 +119,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -regex==2023.5.5 +regex==2023.10.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # transformers @@ -134,10 +128,14 @@ requests==2.31.0 # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # transformers -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir +safetensors==0.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers six==1.16.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -146,32 +144,30 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -tokenizers==0.13.3 +tokenizers==0.14.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # transformers -torch==2.0.1+cu118 - # via - # -r benchmarks/huggingface/requirements.in - # triton -tqdm==4.65.0 +torch==2.1.0+cu118 + # via -r benchmarks/huggingface/requirements.in +tqdm==4.66.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # transformers -transformers==4.29.2 +transformers==4.34.1 # via -r benchmarks/huggingface/requirements.in -triton==2.0.0 +triton==2.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests diff --git a/benchmarks/huggingface/requirements.rocm.txt b/benchmarks/huggingface/requirements.rocm.txt index b8d76be1f..8efd0dc20 100644 --- a/benchmarks/huggingface/requirements.rocm.txt +++ b/benchmarks/huggingface/requirements.rocm.txt @@ -1,28 +1,28 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/huggingface/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-hf.txt benchmarks/huggingface/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/huggingface/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-hf.txt benchmarks/huggingface/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.26.3 +cmake==3.27.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm @@ -34,25 +34,27 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pytorch-triton-rocm # torch # transformers -fsspec==2023.5.0 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -huggingface-hub==0.14.1 +huggingface-hub==0.17.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # tokenizers # transformers idna==3.4 # via @@ -62,15 +64,15 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==16.0.5 +lit==17.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -82,11 +84,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers @@ -98,7 +100,7 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -107,7 +109,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -115,11 +117,11 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -129,7 +131,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -regex==2023.5.5 +regex==2023.10.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers @@ -138,10 +140,14 @@ requests==2.31.0 # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # transformers -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir +safetensors==0.4.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # transformers six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -150,28 +156,28 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -tokenizers==0.13.3 +tokenizers==0.14.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via # -r benchmarks/huggingface/requirements.in # pytorch-triton-rocm -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # transformers -transformers==4.29.2 +transformers==4.34.1 # via -r benchmarks/huggingface/requirements.in -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests diff --git a/benchmarks/rwkv/requirements.cuda.txt b/benchmarks/rwkv/requirements.cuda.txt index 2c0c866e4..a0820efd3 100644 --- a/benchmarks/rwkv/requirements.cuda.txt +++ b/benchmarks/rwkv/requirements.cuda.txt @@ -1,12 +1,12 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/rwkv/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-rwkv.txt benchmarks/rwkv/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/rwkv/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-rwkv.txt benchmarks/rwkv/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 -aiohttp==3.8.4 +aiohttp==3.8.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # fsspec @@ -18,11 +18,11 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -async-timeout==4.0.2 +async-timeout==4.0.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp @@ -30,43 +30,40 @@ attrs==23.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # requests -cmake==3.26.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton codefind==0.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera -deepspeed==0.8.3 +deepspeed==0.11.1 # via -r benchmarks/rwkv/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -frozenlist==1.3.3 +frozenlist==1.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.5.0 +fsspec[http]==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # pytorch-lightning + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -85,19 +82,16 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -lightning-utilities==0.8.0 +lightning-utilities==0.9.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # pytorch-lightning -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton -markdown-it-py==2.2.0 + # torchmetrics +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -114,15 +108,15 @@ multidict==6.0.4 # -c .pin/../.pin/constraints-cuda-torch.txt # aiohttp # yarl -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -ninja==1.11.1 +ninja==1.11.1.1 # via # -r benchmarks/rwkv/requirements.in # deepspeed -numpy==1.24.3 +numpy==1.26.1 # via # -r benchmarks/rwkv/requirements.in # deepspeed @@ -136,14 +130,14 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed # lightning-utilities # pytorch-lightning # torchmetrics -psutil==5.9.5 +psutil==5.9.6 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed @@ -155,11 +149,11 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed -pydantic==1.10.8 +pydantic==1.10.13 # via - # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -169,7 +163,7 @@ pynvml==11.5.0 # voir pytorch-lightning==1.9.5 # via -r benchmarks/rwkv/requirements.in -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf @@ -182,7 +176,7 @@ requests==2.31.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # fsspec -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir @@ -194,27 +188,26 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -torch==2.0.1+cu118 +torch==2.1.0+cu118 # via # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning # torchmetrics - # triton -torchmetrics==0.11.4 +torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # pytorch-lightning -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # deepspeed # pytorch-lightning -triton==2.0.0 +triton==2.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # lightning-utilities @@ -222,7 +215,7 @@ typing-extensions==4.6.2 # pytorch-lightning # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests @@ -230,7 +223,7 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -voir==0.2.9 +voir==0.2.10 # via -r benchmarks/rwkv/requirements.in yarl==1.9.2 # via diff --git a/benchmarks/rwkv/requirements.in b/benchmarks/rwkv/requirements.in index 79d763f72..29a8180d0 100644 --- a/benchmarks/rwkv/requirements.in +++ b/benchmarks/rwkv/requirements.in @@ -3,4 +3,5 @@ torch deepspeed pytorch-lightning<2.0 ninja -voir>=0.2.9,<0.3 +voir>=0.2.10,<0.3 +pydantic<2 diff --git a/benchmarks/rwkv/requirements.rocm.txt b/benchmarks/rwkv/requirements.rocm.txt index 79710b03d..35dd8ff95 100644 --- a/benchmarks/rwkv/requirements.rocm.txt +++ b/benchmarks/rwkv/requirements.rocm.txt @@ -1,12 +1,12 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/rwkv/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-rwkv.txt benchmarks/rwkv/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/rwkv/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-rwkv.txt benchmarks/rwkv/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ -aiohttp==3.8.4 +aiohttp==3.8.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # fsspec @@ -18,11 +18,11 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -async-timeout==4.0.2 +async-timeout==4.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp @@ -30,16 +30,16 @@ attrs==23.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # requests -cmake==3.26.3 +cmake==3.27.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm @@ -47,26 +47,27 @@ codefind==0.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera -deepspeed==0.8.3 +deepspeed==0.11.1 # via -r benchmarks/rwkv/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -frozenlist==1.3.3 +frozenlist==1.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.5.0 +fsspec[http]==2023.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -85,19 +86,20 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lightning-utilities==0.8.0 +lightning-utilities==0.9.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning -lit==16.0.5 + # torchmetrics +lit==17.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -114,15 +116,15 @@ multidict==6.0.4 # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # yarl -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -ninja==1.11.1 +ninja==1.11.1.1 # via # -r benchmarks/rwkv/requirements.in # deepspeed -numpy==1.24.3 +numpy==1.26.1 # via # -r benchmarks/rwkv/requirements.in # deepspeed @@ -136,14 +138,14 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed # lightning-utilities # pytorch-lightning # torchmetrics -psutil==5.9.5 +psutil==5.9.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed @@ -155,11 +157,11 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pydantic==1.10.8 +pydantic==1.10.13 # via - # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -169,11 +171,11 @@ pynvml==11.5.0 # voir pytorch-lightning==1.9.5 # via -r benchmarks/rwkv/requirements.in -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf @@ -186,7 +188,7 @@ requests==2.31.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # fsspec -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -198,23 +200,23 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning # pytorch-triton-rocm # torchmetrics -torchmetrics==0.11.4 +torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed # pytorch-lightning -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # lightning-utilities @@ -222,7 +224,7 @@ typing-extensions==4.6.2 # pytorch-lightning # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests @@ -230,7 +232,7 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.9 +voir==0.2.10 # via -r benchmarks/rwkv/requirements.in yarl==1.9.2 # via diff --git a/benchmarks/stable_baselines3/requirements.cuda.txt b/benchmarks/stable_baselines3/requirements.cuda.txt index a2dcc069c..7af1e0780 100644 --- a/benchmarks/stable_baselines3/requirements.cuda.txt +++ b/benchmarks/stable_baselines3/requirements.cuda.txt @@ -304,7 +304,7 @@ urllib3==1.26.15 # sentry-sdk varname==0.10.0 # via giving -voir==0.2.10 +voir @ git+https://github.com/breuleux/voir.git # via -r benchmarks/stable_baselines3/requirements.in wandb==0.14.0 # via -r benchmarks/stable_baselines3/requirements.in diff --git a/benchmarks/stable_baselines3/requirements.rocm.txt b/benchmarks/stable_baselines3/requirements.rocm.txt index 2096b0592..14b49bd41 100644 --- a/benchmarks/stable_baselines3/requirements.rocm.txt +++ b/benchmarks/stable_baselines3/requirements.rocm.txt @@ -304,7 +304,7 @@ urllib3==1.26.15 # sentry-sdk varname==0.10.0 # via giving -voir==0.2.10 +voir @ git+https://github.com/breuleux/voir.git # via -r benchmarks/stable_baselines3/requirements.in wandb==0.14.0 # via -r benchmarks/stable_baselines3/requirements.in diff --git a/benchmarks/stargan/requirements.cuda.txt b/benchmarks/stargan/requirements.cuda.txt index 16485fb78..b66e81e63 100644 --- a/benchmarks/stargan/requirements.cuda.txt +++ b/benchmarks/stargan/requirements.cuda.txt @@ -1,8 +1,8 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/stargan/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-stargan.txt benchmarks/stargan/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/stargan/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-stargan.txt benchmarks/stargan/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -cmake==3.26.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton codefind==0.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -34,11 +30,15 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton +fsspec==2023.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -52,15 +52,11 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -72,11 +68,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -r benchmarks/stargan/requirements.in # torchvision @@ -88,7 +84,7 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision @@ -96,7 +92,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -104,7 +100,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf @@ -116,7 +112,7 @@ requests==2.31.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir @@ -128,23 +124,22 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -torch==2.0.1+cu118 +torch==2.1.0+cu118 # via # -r benchmarks/stargan/requirements.in # torchvision - # triton -torchvision==0.15.2+cu118 +torchvision==0.16.0+cu118 # via -r benchmarks/stargan/requirements.in -triton==2.0.0 +triton==2.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests diff --git a/benchmarks/stargan/requirements.rocm.txt b/benchmarks/stargan/requirements.rocm.txt index 594c94949..b4eb2f7e1 100644 --- a/benchmarks/stargan/requirements.rocm.txt +++ b/benchmarks/stargan/requirements.rocm.txt @@ -1,28 +1,28 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/stargan/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-stargan.txt benchmarks/stargan/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/stargan/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-stargan.txt benchmarks/stargan/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.26.3 +cmake==3.27.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm @@ -34,11 +34,15 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch +fsspec==2023.1.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -52,15 +56,15 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==16.0.5 +lit==17.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -72,11 +76,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -r benchmarks/stargan/requirements.in # torchvision @@ -88,7 +92,7 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -96,7 +100,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -104,11 +108,11 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf @@ -120,7 +124,7 @@ requests==2.31.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,19 +136,19 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via # -r benchmarks/stargan/requirements.in # pytorch-triton-rocm # torchvision -torchvision==0.15.2+rocm5.4.2 +torchvision==0.16.0+rocm5.6 # via -r benchmarks/stargan/requirements.in -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests diff --git a/benchmarks/super-slomo/requirements.cuda.txt b/benchmarks/super-slomo/requirements.cuda.txt index 40caddc4d..a97179600 100644 --- a/benchmarks/super-slomo/requirements.cuda.txt +++ b/benchmarks/super-slomo/requirements.cuda.txt @@ -1,8 +1,8 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/super-slomo/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-super-slomo.txt benchmarks/super-slomo/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/super-slomo/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-super-slomo.txt benchmarks/super-slomo/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -cmake==3.26.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton codefind==0.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -34,11 +30,15 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton +fsspec==2023.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -52,15 +52,11 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -72,11 +68,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -r benchmarks/super-slomo/requirements.in # opencv-python @@ -85,13 +81,13 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -opencv-python==4.7.0.72 +opencv-python==4.8.1.78 # via -r benchmarks/super-slomo/requirements.in ovld==0.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision @@ -99,7 +95,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -107,7 +103,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf @@ -119,7 +115,7 @@ requests==2.31.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir @@ -131,25 +127,24 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -torch==2.0.1+cu118 +torch==2.1.0+cu118 # via # -r benchmarks/super-slomo/requirements.in # torchvision - # triton -torchvision==0.15.2+cu118 +torchvision==0.16.0+cu118 # via -r benchmarks/super-slomo/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via -r benchmarks/super-slomo/requirements.in -triton==2.0.0 +triton==2.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests diff --git a/benchmarks/super-slomo/requirements.rocm.txt b/benchmarks/super-slomo/requirements.rocm.txt index a3557abc5..b96de3447 100644 --- a/benchmarks/super-slomo/requirements.rocm.txt +++ b/benchmarks/super-slomo/requirements.rocm.txt @@ -1,28 +1,28 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/super-slomo/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-super-slomo.txt benchmarks/super-slomo/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/super-slomo/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-super-slomo.txt benchmarks/super-slomo/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.26.3 +cmake==3.27.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm @@ -34,11 +34,15 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch +fsspec==2023.1.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -52,15 +56,15 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==16.0.5 +lit==17.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -72,11 +76,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -r benchmarks/super-slomo/requirements.in # opencv-python @@ -85,13 +89,13 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -opencv-python==4.7.0.72 +opencv-python==4.8.1.78 # via -r benchmarks/super-slomo/requirements.in ovld==0.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -99,7 +103,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -107,11 +111,11 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf @@ -123,7 +127,7 @@ requests==2.31.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -135,21 +139,21 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via # -r benchmarks/super-slomo/requirements.in # pytorch-triton-rocm # torchvision -torchvision==0.15.2+rocm5.4.2 +torchvision==0.16.0+rocm5.6 # via -r benchmarks/super-slomo/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via -r benchmarks/super-slomo/requirements.in -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests diff --git a/benchmarks/timm/requirements.cuda.txt b/benchmarks/timm/requirements.cuda.txt index e23485257..9d276bd86 100644 --- a/benchmarks/timm/requirements.cuda.txt +++ b/benchmarks/timm/requirements.cuda.txt @@ -1,8 +1,8 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/timm/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-timm.txt benchmarks/timm/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/timm/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-timm.txt benchmarks/timm/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -cmake==3.26.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton codefind==0.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -34,22 +30,23 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # torch # triton -fsspec==2023.5.0 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -huggingface-hub==0.14.1 +huggingface-hub==0.17.3 # via -r benchmarks/timm/requirements.in idna==3.4 # via @@ -59,15 +56,11 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -79,11 +72,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision @@ -95,11 +88,11 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision @@ -107,7 +100,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -115,7 +108,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -r benchmarks/timm/requirements.in # huggingface-hub @@ -129,11 +122,11 @@ requests==2.31.0 # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # torchvision -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -safetensors==0.3.1 +safetensors==0.4.0 # via -r benchmarks/timm/requirements.in six==1.16.0 # via @@ -143,28 +136,27 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -torch==2.0.1+cu118 +torch==2.1.0+cu118 # via # -r benchmarks/timm/requirements.in # torchvision - # triton -torchvision==0.15.2+cu118 +torchvision==0.16.0+cu118 # via -r benchmarks/timm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub -triton==2.0.0 +triton==2.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # huggingface-hub # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests diff --git a/benchmarks/timm/requirements.rocm.txt b/benchmarks/timm/requirements.rocm.txt index 12864565b..fb3c00832 100644 --- a/benchmarks/timm/requirements.rocm.txt +++ b/benchmarks/timm/requirements.rocm.txt @@ -1,28 +1,28 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/timm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-timm.txt benchmarks/timm/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/timm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-timm.txt benchmarks/timm/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.26.3 +cmake==3.27.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm @@ -34,22 +34,23 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pytorch-triton-rocm # torch -fsspec==2023.5.0 +fsspec==2023.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -huggingface-hub==0.14.1 +huggingface-hub==0.17.3 # via -r benchmarks/timm/requirements.in idna==3.4 # via @@ -59,15 +60,15 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==16.0.5 +lit==17.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -79,11 +80,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -95,11 +96,11 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.1 +packaging==23.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -107,7 +108,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -115,11 +116,11 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -r benchmarks/timm/requirements.in # huggingface-hub @@ -133,11 +134,11 @@ requests==2.31.0 # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # torchvision -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -safetensors==0.3.1 +safetensors==0.4.0 # via -r benchmarks/timm/requirements.in six==1.16.0 # via @@ -147,24 +148,24 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via # -r benchmarks/timm/requirements.in # pytorch-triton-rocm # torchvision -torchvision==0.15.2+rocm5.4.2 +torchvision==0.16.0+rocm5.6 # via -r benchmarks/timm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests diff --git a/benchmarks/torchvision/main.py b/benchmarks/torchvision/main.py index a52cbbec0..b0a7cf847 100644 --- a/benchmarks/torchvision/main.py +++ b/benchmarks/torchvision/main.py @@ -165,7 +165,10 @@ def main(): if data_directory: args.data = os.path.join(data_directory, "FakeImageNet") - use_cuda = not args.no_cuda and torch.cuda.is_available() + if not args.no_cuda: + assert torch.cuda.is_available(), "Why is CUDA not available" + + use_cuda = not args.no_cuda torch.manual_seed(args.seed) if use_cuda: diff --git a/benchmarks/torchvision/requirements.cuda.txt b/benchmarks/torchvision/requirements.cuda.txt index aa60bdcfb..b2770329c 100644 --- a/benchmarks/torchvision/requirements.cuda.txt +++ b/benchmarks/torchvision/requirements.cuda.txt @@ -1,8 +1,8 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/torchvision/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests -cmake==3.26.3 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton codefind==0.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -34,11 +30,15 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton +fsspec==2023.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt @@ -52,15 +52,11 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-cuda-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 @@ -72,11 +68,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision @@ -88,7 +84,7 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision @@ -96,7 +92,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # rich @@ -104,7 +100,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # omegaconf @@ -116,7 +112,7 @@ requests==2.31.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torchvision -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # voir @@ -128,25 +124,24 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -torch==2.0.1+cu118 +torch==2.1.0+cu118 # via # -r benchmarks/torchvision/requirements.in # torchvision - # triton -torchvision==0.15.2+cu118 +torchvision==0.16.0+cu118 # via -r benchmarks/torchvision/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via -r benchmarks/torchvision/requirements.in -triton==2.0.0 +triton==2.1.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-cuda-torch.txt # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-cuda-torch.txt # requests diff --git a/benchmarks/torchvision/requirements.rocm.txt b/benchmarks/torchvision/requirements.rocm.txt index 09ebf1695..03376298c 100644 --- a/benchmarks/torchvision/requirements.rocm.txt +++ b/benchmarks/torchvision/requirements.rocm.txt @@ -1,28 +1,28 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.5.7 +certifi==2023.7.22 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.26.3 +cmake==3.27.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm @@ -34,11 +34,15 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.12.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch +fsspec==2023.1.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -52,15 +56,15 @@ jinja2==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==16.0.5 +lit==17.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -72,11 +76,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.1 +networkx==3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.24.3 +numpy==1.26.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,7 +92,7 @@ ovld==0.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==9.5.0 +pillow==10.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -96,7 +100,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.15.1 +pygments==2.16.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -104,11 +108,11 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf @@ -120,7 +124,7 @@ requests==2.31.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision -rich==13.3.5 +rich==13.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,21 +136,21 @@ sympy==1.12 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.0.1+rocm5.4.2 +torch==2.1.0+rocm5.6 # via # -r benchmarks/torchvision/requirements.in # pytorch-triton-rocm # torchvision -torchvision==0.15.2+rocm5.4.2 +torchvision==0.16.0+rocm5.6 # via -r benchmarks/torchvision/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via -r benchmarks/torchvision/requirements.in -typing-extensions==4.6.2 +typing-extensions==4.8.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # reactivex # torch -urllib3==1.26.16 +urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests diff --git a/constraints/cuda.txt b/constraints/cuda.txt index 68242da32..2f4bcb7f8 100644 --- a/constraints/cuda.txt +++ b/constraints/cuda.txt @@ -1,2 +1 @@ ---extra-index-url https://download.pytorch.org/whl/cu118 -deepspeed==0.8.3 +--extra-index-url https://download.pytorch.org/whl/cu118 \ No newline at end of file diff --git a/constraints/rocm.txt b/constraints/rocm.txt index 8d434fa36..f76663f04 100644 --- a/constraints/rocm.txt +++ b/constraints/rocm.txt @@ -1,2 +1 @@ ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ -deepspeed==0.8.3 +--extra-index-url https://download.pytorch.org/whl/rocm5.6/ \ No newline at end of file diff --git a/milabench/_version.py b/milabench/_version.py index 5a191cf45..8fc180822 100644 --- a/milabench/_version.py +++ b/milabench/_version.py @@ -1,5 +1,5 @@ """This file is generated, do not modify""" -__tag__ = "cce4c99" -__commit__ = "cce4c994559481334f9c15b01caa2d4e7b0cf82a" -__date__ = "2023-07-04 17:33:49 -0400" +__tag__ = "v0.0.6-43-g89f56f6" +__commit__ = "89f56f670db0f22880d057262a320c935c217d77" +__date__ = "2023-10-19 19:29:36 -0400" diff --git a/milabench/cli.py b/milabench/cli.py index d163114ff..08294852b 100644 --- a/milabench/cli.py +++ b/milabench/cli.py @@ -1,6 +1,7 @@ import json import os import re +import io import runpy import shutil import subprocess @@ -33,6 +34,7 @@ from .report import make_report from .slurm import expand_node_list from .summary import aggregate, make_summary +from .schedule import launch_milabench, post_comment_on_pr def main(argv=None): @@ -253,6 +255,30 @@ def is_selected(defn): ) + +def _parse_report(pth): + with pth.open() as f: + lines = f.readlines() + data = [] + good_lines = 0 + bad_lines = 0 + + for line in lines: + try: + data.append(json.loads(line)) + good_lines += 1 + except Exception: + import traceback + print(f"Could not parse line inside {pth}\n\t- {line}") + traceback.print_exc() + bad_lines += 1 + + if good_lines == 0: + raise RuntimeError(f"Unknow format for file {pth}") + + return data + + def _read_reports(*runs): all_data = {} for folder in runs: @@ -261,17 +287,8 @@ def _read_reports(*runs): if not file.endswith(".data"): continue pth = XPath(parent) / file - with pth.open() as f: - lines = f.readlines() - try: - data = [json.loads(line) for line in lines] - except Exception: - import traceback - - print(f"Could not parse line inside {pth}\n\t- {line}") - traceback.print_exc() - else: - all_data[str(pth)] = data + all_data[str(pth)] = _parse_report(pth) + return all_data @@ -353,7 +370,7 @@ def run(): report: Option & bool = True # Which type of dashboard to show (short, long, or no) - dash: Option & str = os.environ.get("MILABENCH_DASH", "long") + dash: Option & str = os.getenv("MILABENCH_DASH", "long") noterm: Option & bool = os.getenv("MILABENCH_NOTERM", "0") == "1" @@ -379,8 +396,9 @@ def run(): # Terminal Formatter slows down the dashboard, # if lots of info needs to be printed # in particular rwkv - TerminalFormatter() if not noterm else None, - dash_class and dash_class(), + # TerminalFormatter() if not noterm else None, + # dash_class and dash_class(), + TerminalFormatter(), TextReporter("stdout"), TextReporter("stderr"), DataReporter(), @@ -395,11 +413,14 @@ def run(): compare_gpus = False html = None price = None - + reports = None if runs: reports = _read_reports(*runs) + assert len(reports) != 0, "No reports found" + summary = make_summary(reports.values()) + assert len(summary) != 0, "No summaries" make_report( summary, @@ -734,201 +755,81 @@ def publish(): backend = SQLAlchemy(uri, meta_override=meta) publish_archived_run(backend, folder) - def container(): - """Build a container image (might not work properly at the moment).""" - - # Configuration file - # [positional] - config_file: Option & str = None - - config = _get_multipack(config, return_config=True) - config_file = XPath(config["defaults"]["config_file"]) - config_base = XPath(config["defaults"]["config_base"]) - benchmarks = config["benchmarks"] - - # The container type to create - type: Option & str = None - - # Include the dataset in the image - include_data: Option & bool = False - - # Optional path to copy build dir to, instead of building the image. - # This directory must not exist and will be created. - output_dir: Option & str = None - - # File in which to generate the SIF image (Singularity). - # Defaults to milabench.sif. - # [alias: -o] - output_file: Option & str = None - - # Optional python version to use for the image, ignored for - # conda-based benchmarks. Can be specified as any of - # ('3', '3.9', '3.9.2') - python_version: Option & str = "3.9" - - # Milabench source to clone from - milabench: Option & str = "v2" - - # The tag for the generated container - tag: Option & str = None - - if type not in ["docker", "singularity"]: - sys.exit(f"Unsupported type {type}") - - with tempfile.TemporaryDirectory() as base: - root = XPath(base) - - common_base = config_base - - # Figure out common base between the benchmark config and all - # the benchmarks. - for defn in benchmarks.values(): - pack = XPath(defn["definition"]).expanduser() - while not pack.is_relative_to(common_base): - common_base = common_base.parent - - def _transfer(pth): - dest = root / pth.relative_to(common_base) - shutil.copytree(pth, dest, dirs_exist_ok=True) - - for defn in benchmarks.values(): - _transfer(XPath(defn["definition"])) - - _transfer(config_base) - - # We check all configs since they may not have all the same setting - use_conda = any( - defn["venv"]["type"] == "conda" for defn in benchmarks.values() - ) - - if "//" not in milabench: - milabench = ( - f"git+https://github.com/mila-iqia/milabench.git@{milabench}" - ) - - if type == "docker": - if output_file is not None: - sys.exit("Error: --output-file only valid with Singularity") - tag = tag or "milabench" - with (root / "Dockerfile").open("w") as f: - f.write( - dockerfile_template( - milabench_req=milabench, - include_data=include_data, - use_conda=use_conda, - python_version=python_version, - config_file=config_file.relative_to(common_base), - ) - ) - if output_dir: - root.copy(output_dir) - else: - subprocess.check_call(["docker", "build", ".", "-t", tag], cwd=root) - - elif type == "singularity": - if tag is not None: - sys.exit("Error: --tag only valid with Docker") - output_file = output_file or "milabench.sif" - - with (root / "milabench.def").open("w") as f: - f.write( - singularitydef_template( - milabench_req=milabench, - include_data=include_data, - use_conda=use_conda, - python_version=python_version, - config_file=config_file.relative_to(common_base), - ) - ) - if output_dir: - root.copy(output_dir) - else: - user = os.environ["USER"] - filename = str(XPath(output_file).absolute()) - singularity = subprocess.check_output( - ["which", "singularity"] - ).strip() - subprocess.check_call( - ["sudo", singularity, "build", filename, "milabench.def"], - cwd=root, - ) - subprocess.check_call(["sudo", "chown", f"{user}:{user}", filename]) - - -def dockerfile_template( - milabench_req, include_data, use_conda, python_version, config_file -): - conda_clean = "conda clean -a" if use_conda else "echo" - return f""" -FROM { 'continuumio/miniconda3' if use_conda else f'python:{python_version}-slim' } - -RUN apt-get update && apt-get install --no-install-suggests --no-install-recommends -y \ - git \ - wget \ - patch \ - && apt-get clean - -RUN mkdir /bench && mkdir /base -ENV MILABENCH_BASE /base -# This is to signal to milabench to use that as fallback -ENV VIRTUAL_ENV /base/venv/_ -ENV MILABENCH_CONFIG /bench/{ config_file } -ENV HEADLESS 1 -WORKDIR /base - -RUN echo '{ milabench_req }' > /version.txt - -COPY / /bench + def schedule(): + """Launch a slurm job to run milabench""" + # milabench schedule --sync -- --select resnet50 + + # tail -f on the slurm job + sync: Option & bool = False + + # Print the command and return without running it + dry: Option & bool = False + + # pip arguments + # [remainder] + args: Option = [] + + launch_milabench( + args, + sbatch_args=None, + dry=dry, + sync=sync + ) + + def write_report_to_pr(): + remote: str & Option + + branch: str & Option + + base: Option & str = os.getenv("MILABENCH_BASE", None) + + config: Option & str = os.getenv("MILABENCH_CONFIG", None) + + token: str & Option = os.getenv("MILABENCH_GITHUB_PAT") + + assert base is not None + + runfolder = os.path.join(base, "runs") + + def filter(folder): + for f in ('install', 'prepare'): + if f in folder: + return False + return True + + runs = [] + for folder in os.listdir(runfolder): + if filter(folder): + runs.append(os.path.join(runfolder, folder)) + + report = _short_make_report(runs, config) + + post_comment_on_pr( + remote, + branch, + "```\n" + report + "\n```", + token + ) -RUN pip install -U pip && \ - pip install -r /version.txt && \ - milabench install && \ - { conda_clean } && \ - pip cache purge -{ 'RUN milabench prepare' if include_data else '' } +def _short_make_report(runs, config): + reports = None -CMD ["milabench", "run"] -""" + if runs: + reports = _read_reports(*runs) + summary = make_summary(reports.values()) + if config: + config = _get_multipack(config, return_config=True) -def singularitydef_template( - milabench_req, include_data, use_conda, python_version, config_file -): - conda_clean = "conda clean -a" if use_conda else "echo" - return f"""\ -BootStrap: docker -From: { 'continuumio/miniconda3' if use_conda else f'python:{python_version}-slim' } - -%files - . /bench - -%environment - export MILABENCH_BASE=/base - export MILABENCH_CONFIG=/bench/{ config_file } - export HEADLESS=1 - -%post - export MILABENCH_BASE=/base - export MILABENCH_CONFIG=/bench/{ config_file } - export HEADLESS=1 - - apt-get update && apt-get install --no-install-suggests --no-install-recommends -y git wget patch - apt-get clean - - mkdir /base - cd /bench - - echo '{ milabench_req }' > /version.txt - pip install -U pip && \ - pip install -r /version.txt && \ - milabench install && \ - { conda_clean } && \ - pip cache purge -{ ' milabench prepare' if include_data else '' } - - chmod -R o+rwx /base /bench - -%runscript - milabench run -""" + stream = io.StringIO() + + make_report( + summary, + weights=config, + stream=stream, + sources=runs, + errdata=reports and _error_report(reports), + ) + + return stream.getvalue() \ No newline at end of file diff --git a/milabench/config.py b/milabench/config.py index da29da294..88053ca20 100644 --- a/milabench/config.py +++ b/milabench/config.py @@ -101,6 +101,26 @@ def get_remote_ip(): return set(result) +def _resolve_ip(ip): + # Resolve the IP + try: + hostname, aliaslist, ipaddrlist = socket.gethostbyaddr(ip) + lazy_raise = None + except socket.gaierror as err: + # Get Addr Info (GAI) Error + # + # When we are connecting to a node through a ssh proxy jump + # the node IPs/Hostnames are not available until we reach + # the first node inside the cluster + # + hostname = ip + aliaslist = [] + ipaddrlist = [] + lazy_raise = err + + return hostname, aliaslist, ipaddrlist, lazy_raise + + def resolve_addresses(nodes): # Note: it is possible for self to be none # if we are running milabench on a node that is not part of the system @@ -111,26 +131,20 @@ def resolve_addresses(nodes): ip_list = get_remote_ip() for node in nodes: - # Resolve the IP - try: - hostname, aliaslist, ipaddrlist = socket.gethostbyaddr(node["ip"]) - - except socket.gaierror as err: - # Get Addr Info (GAI) Error - # - # When we are connecting to a node through a ssh proxy jump - # the node IPs/Hostnames are not available until we reach - # the first node inside the cluster - # - hostname = node["ip"] - aliaslist = [] - ipaddrlist = [] - - lazy_raise = err + hostname, aliaslist, ipaddrlist, lazy_raise = _resolve_ip(node["ip"]) node["hostname"] = hostname node["aliaslist"] = aliaslist node["ipaddrlist"] = ipaddrlist + + if hostname.endswith(".server.mila.quebec.server.mila.quebec"): + print() + print("Hostname was extra long for no reason") + print(hostname, socket.gethostname()) + print() + + # why is this happening + hostname = hostname[:-len(".server.mila.quebec")] is_local = ( ("127.0.0.1" in ipaddrlist) diff --git a/milabench/executors.py b/milabench/executors.py index 5e5f9a3d8..41d308919 100644 --- a/milabench/executors.py +++ b/milabench/executors.py @@ -425,7 +425,7 @@ def _find_node_config(self) -> Dict: def is_local(self): localnode = self.pack.config["system"]["self"] - + # self is none; the node we are currently # on is not part of the system; we are running # milabench remotely, sending remote commands to @@ -660,7 +660,7 @@ def _argv(self, **_) -> List: ngpu = len(get_gpu_info()["gpus"].values()) nproc = ngpu * num_machines - assert nproc > 0 + assert nproc > 0, f"nproc: {nproc} num_machines: {num_machines} ngpu: {ngpu}" deepspeed_argv = ( [ diff --git a/milabench/pack.py b/milabench/pack.py index 2656fec55..a1bf5c3cb 100644 --- a/milabench/pack.py +++ b/milabench/pack.py @@ -376,7 +376,7 @@ async def pin( ivar = self.config.get("install_variant", None) if ivar == "unpinned": raise Exception("Cannot pin the 'unpinned' variant.") - assert self.phase == "pin" + # assert self.phase == "pin" for base_reqs, reqs in self.requirements_map().items(): if not base_reqs.exists(): raise FileNotFoundError( diff --git a/milabench/report.py b/milabench/report.py index 1c0c63f56..f8de43abf 100644 --- a/milabench/report.py +++ b/milabench/report.py @@ -41,15 +41,19 @@ def _make_row(summary, compare, weights): # Sum of all the GPU performance # to get the overall perf of the whole machine - acc = 0 - for _, metrics in summary["per_gpu"].items(): - acc += metrics[metric] + + if "per_gpu" in summary: + acc = 0 + for _, metrics in summary["per_gpu"].items(): + acc += metrics[metric] + else: + acc = row["perf"] success_ratio = 1 - row["fail"] / row["n"] score = (acc if acc > 0 else row["perf"]) * success_ratio row["score"] = score - row["weight"] = weights.get("weight", summary["weight"]) + row["weight"] = weights.get("weight", summary.get("weight", 0)) # ---- return row @@ -219,16 +223,17 @@ def make_report( sources=None, errdata=None, weights=None, + stream=sys.stdout ): if weights is None: weights = dict() df = make_dataframe(summary, compare, weights) - + # Reorder columns df = df[sorted(df.columns, key=lambda k: columns_order.get(k, 0))] - - out = Outputter(stdout=sys.stdout, html=html) + + out = Outputter(stdout=stream, html=html) if sources: if isinstance(sources, str): diff --git a/milabench/schedule.py b/milabench/schedule.py new file mode 100644 index 000000000..c572e0e38 --- /dev/null +++ b/milabench/schedule.py @@ -0,0 +1,212 @@ + +from dataclasses import dataclass +import re +import importlib_resources +import subprocess +import requests +import os + + +def popen(cmd, callback=None): + def println(line): + print(line, end="") + + if callback is None: + callback=println + + with subprocess.Popen( + cmd, + stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, + text=True, + shell=False, + ) as process: + def readoutput(): + process.stdout.flush() + line = process.stdout.readline() + + if callback: + callback(line) + + try: + while process.poll() is None: + readoutput() + + readoutput() + return 0 + + except KeyboardInterrupt: + print("Stopping due to user interrupt") + process.kill() + return -1 + +def sbatch(args, sync=False, tags=None, **kwargs): + jobid_regex = re.compile(r"Submitted batch job (?P[0-9]*)") + jobid = None + + def readline(line): + nonlocal jobid + + if match := jobid_regex.match(line): + data = match.groupdict() + jobid = data["jobid"] + + print(line, end="") + + code = popen(['sbatch'] + args, readline) + + if jobid is not None and sync: + try: + subprocess.run(["touch", f"slurm-{jobid}.out"]) + subprocess.run(["tail", "-f", f"slurm-{jobid}.out"]) + except KeyboardInterrupt: + pass + + return code, jobid + + +def shell(cmd): + return subprocess.check_output( + cmd.split(" "), + stderr=subprocess.STDOUT, + text=True + ).strip() + + +class SlurmBatchOptions: + pass + + +@dataclass +class SetupOptions: + branch: str = "master" + origin: str = "https://github.com/mila-iqia/milabench.git" + config: str = "milabench/config/standard.yaml" + env: str = "./env" + python: str = "3.9" + + def deduce_remote(self, current_branch): + prefix = "refs/heads/" + + # Fetch all remotes + remotes = shell("git remote").splitlines() + possible_remotes = [] + + # Find remotes that have our branch + for remote in remotes: + branches = shell(f"git ls-remote --heads {remote}").splitlines() + + for branch in branches: + _, name = branch.split('\t') + name = name[len(prefix):] + + if current_branch == name: + possible_remotes.append(remote) + + if len(possible_remotes) == 1: + return possible_remotes[0] + + raise RuntimeError(f"Multiple suitable remotes found {possible_remotes}") + + def deduce_from_repository(self, remote=None): + self.branch = shell("git rev-parse --abbrev-ref HEAD") + + if remote is None: + remote = self.deduce_remote(self.branch) + + self.origin = shell(f"git remote get-url {remote}") + + def arguments(self): + return [ + "-b", self.branch, + "-o", self.origin, + "-c", self.config, + "-e", self.env, + "-p", self.python, + ] + + +def launch_milabench(args, sbatch_args=None, dry: bool = False, sync: bool = False): + sbatch_script = importlib_resources.files(__name__) / "scripts" / "milabench_run.bash" + sbatch_script = str(sbatch_script) + + if sbatch_args is None: + sbatch_args = [ + "--ntasks=1", + "--gpus-per-task=rtx8000:1", + "--cpus-per-task=4", + "--time=01:30:00", + "--ntasks-per-node=1", + "--mem=64G" + ] + + script_args = SetupOptions() + script_args.deduce_from_repository() + script_args = script_args.arguments() + + cmd = sbatch_args + [sbatch_script] + script_args + args + + if dry: + print("sbatch " + ' '.join(cmd)) + code = 0 + else: + code, _ = sbatch(cmd, sync=sync, tags=None) + + return code + + +def get_remote_owner(remote): + sshremote = re.compile(r"git@[A-Za-z]*\.[A-Za-z]*:(?P[A-Za-z\-.0-9]*)\/([A-Za-z]*).([A-Za-z]*)") + httpsremote = re.compile(r"https:\/\/[A-Za-z]*\.[A-Za-z]*\/(?P[A-Za-z\-.0-9]*)\/([A-Za-z]*).([A-Za-z]*)") + + patterns = [sshremote, httpsremote] + + for pat in patterns: + if match := sshremote.match(remote): + results = match.groupdict() + return results['owner'] + + return None + + +def post_comment_on_pr(remote, branch, comment, access_token=None): + owner = get_remote_owner(remote) + assert owner is not None, "Remote owner not found" + + if access_token is None: + access_token = os.getenv("MILABENCH_GITHUB_PAT") + + url = "https://api.github.com/repos/mila-iqia/milabench/pulls" + + response = requests.get(url, params={"head": f"{owner}:{branch}"}) + + if response.status_code != 200: + raise RuntimeError(response) + + pull_requests = response.json() + + if not pull_requests: + raise RuntimeError("No matching pull requests found.") + + assert len(pull_requests) == 1, "Multiple PR found" + + pr = pull_requests[0] + post_url = pr["_links"]["comments"]["href"] + + data = { + "body": comment, + } + + headers = { + "Authorization": f"Bearer {access_token}", + "Accept": "application/vnd.github.v3+json" + } + + response = requests.post( + post_url, + json=data, + headers=headers + ) + + if response.status_code != 201: + raise RuntimeError(response, response.json()) diff --git a/milabench/scripts/milabench_pin.bash b/milabench/scripts/milabench_pin.bash new file mode 100644 index 000000000..c2f7ad399 --- /dev/null +++ b/milabench/scripts/milabench_pin.bash @@ -0,0 +1,15 @@ +#!/bin/bash + +# CPU only +# 16Gb + + + +MILABENCH_GPU_ARCH=cuda milabench pin --config config/standard.yaml --from-scratch --base /tmp +MILABENCH_GPU_ARCH=rocm milabench pin --config config/standard.yaml --from-scratch --base /tmp + + +cd $SLURM_TMPDIR/milabench +git add --all +git commit -m "milabench pin" +git push $ORIGIN $BRANCH \ No newline at end of file diff --git a/milabench/scripts/milabench_run.bash b/milabench/scripts/milabench_run.bash new file mode 100755 index 000000000..cf502cbf4 --- /dev/null +++ b/milabench/scripts/milabench_run.bash @@ -0,0 +1,136 @@ +#!/bin/bash + +function usage() { + echo "Usage: $0 [-m] [-p]" + echo " -h Display this help message." + echo " -b arch GPU arch (default: cuda)" + echo " -b BRANCH Branch to checkout (default: master)" + echo " -o ORIGIN Origin to use (default: github/mila/milabench)" + echo " -c CONFIG Configuration (default: milabench/config/standard.yaml)" + echo " -e ENV Environment (default: ./env)" + echo " -p PYTHON Python version (default: 3.9)" + echo " ARGUMENT Any additional argument you want to process." + exit 1 +} + +ARCH="cuda" +PYTHON="3.9" +BRANCH="master" +ORIGIN="https://github.com/mila-iqia/milabench.git" +CONFIG="$SLURM_TMPDIR/milabench/config/standard.yaml" +BASE="$SLURM_TMPDIR/base" +ENV="./env" +REMAINING_ARGS="" + +while getopts ":hm:p:e:b:o:c:" opt; do + case $opt in + h) + usage + ;; + p) + PYTHON="$OPTARG" + ;; + b) + BRANCH="$OPTARG" + ;; + o) + ORIGIN="$OPTARG" + ;; + c) + CONFIG="$OPTARG" + ;; + e) + ENV="$OPTARG" + ;; + a) + ARCH="$OPTARG" + ;; + :) + echo "Option -$OPTARG requires an argument." >&2 + usage + ;; + esac +done + +shift "$((OPTIND-1))" +REMAINING_ARGS="$@" + +echo " PYTHON: $PYTHON" +echo " branch: $BRANCH" +echo " origin: $ORIGIN" +echo " config: $CONFIG" +echo " env: $ENV" +echo " args: $REMAINING_ARGS" +# +# Fix problem with conda saying it is not "init properly" +# +CONDA_EXEC="$(which conda)" +CONDA_BASE=$(dirname $CONDA_EXEC) +source $CONDA_BASE/../etc/profile.d/conda.sh + +if [ -e $HOME/.credentials.env ]; then + source $HOME/.credentials.env +fi + +cd $SLURM_TMPDIR +# +# Create a new environment +# +if [ ! -d "$ENV" ] && [ "$ENV" != "base" ] && [ ! -d "$CONDA_ENVS/$ENV" ]; then + conda create --prefix $ENV python=$PYTHON -y +fi +conda activate $ENV + +export HF_HOME=$BASE/cache +export HF_DATASETS_CACHE=$BASE/cache +export TORCH_HOME=$BASE/cache +export XDG_CACHE_HOME=$BASE/cache +export MILABENCH_GPU_ARCH=$ARCH + +export MILABENCH_DASH=no +export PYTHONUNBUFFERED=1 +export MILABENCH_BASE=$BASE +export MILABENCH_CONFIG=$CONFIG +# +# Fetch the repo +# +git clone --single-branch --depth 1 -b $BRANCH $ORIGIN +python -m pip install -e ./milabench + +SYSTEM="$SLURM_TMPDIR/system.yaml" + +echo "" +echo "System" +echo "------" + +milabench slurm_system +milabench slurm_system > $SYSTEM + +module load gcc/9.3.0 +module load cuda/11.8 + +echo "" +echo "Install" +echo "-------" +milabench install --config $CONFIG --system $SYSTEM --base $BASE $REMAINING_ARGS + + +echo "" +echo "Prepare" +echo "-------" +milabench prepare --config $CONFIG --system $SYSTEM --base $BASE $REMAINING_ARGS + +echo "" +echo "Run" +echo "---" +milabench run --config $CONFIG --system $SYSTEM --base $BASE $REMAINING_ARGS + +echo "" +echo "Report" +echo "------" + +milabench write_report_to_pr --remote $ORIGIN --branch $BRANCH --config $CONFIG + +echo "----" +echo "Done after $SECONDS" +echo "" diff --git a/poetry.lock b/poetry.lock index 6c0c15f1d..942bff025 100644 --- a/poetry.lock +++ b/poetry.lock @@ -1,10 +1,9 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.6.1 and should not be changed by hand. [[package]] name = "alabaster" version = "0.7.13" description = "A configurable sidebar-enabled Sphinx theme" -category = "dev" optional = false python-versions = ">=3.6" files = [ @@ -16,7 +15,6 @@ files = [ name = "ansicon" version = "1.89.0" description = "Python wrapper for loading Jason Hood's ANSICON" -category = "main" optional = false python-versions = "*" files = [ @@ -28,7 +26,6 @@ files = [ name = "antlr4-python3-runtime" version = "4.9.3" description = "ANTLR 4.9.3 runtime for Python 3.7" -category = "main" optional = false python-versions = "*" files = [ @@ -39,7 +36,6 @@ files = [ name = "argcomplete" version = "1.12.3" description = "Bash tab completion for argparse" -category = "main" optional = false python-versions = "*" files = [ @@ -54,7 +50,6 @@ test = ["coverage", "flake8", "pexpect", "wheel"] name = "asttokens" version = "2.2.1" description = "Annotate AST trees with source code positions" -category = "main" optional = false python-versions = "*" files = [ @@ -72,7 +67,6 @@ test = ["astroid", "pytest"] name = "atomicwrites" version = "1.4.1" description = "Atomic file writes." -category = "dev" optional = false python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" files = [ @@ -83,7 +77,6 @@ files = [ name = "attrs" version = "23.1.0" description = "Classes Without Boilerplate" -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -102,7 +95,6 @@ tests-no-zope = ["cloudpickle", "hypothesis", "mypy (>=1.1.1)", "pympler", "pyte name = "babel" version = "2.12.1" description = "Internationalization utilities" -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -117,7 +109,6 @@ pytz = {version = ">=2015.7", markers = "python_version < \"3.9\""} name = "black" version = "23.3.0" description = "The uncompromising code formatter." -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -167,7 +158,6 @@ uvloop = ["uvloop (>=0.15.2)"] name = "blessed" version = "1.20.0" description = "Easy, practical library for making terminal apps, by providing an elegant, well-documented interface to Colors, Keyboard input, and screen Positioning capabilities." -category = "main" optional = false python-versions = ">=2.7" files = [ @@ -184,7 +174,6 @@ wcwidth = ">=0.1.4" name = "build" version = "0.10.0" description = "A simple, correct Python build frontend" -category = "main" optional = false python-versions = ">= 3.7" files = [ @@ -208,7 +197,6 @@ virtualenv = ["virtualenv (>=20.0.35)"] name = "certifi" version = "2023.5.7" description = "Python package for providing Mozilla's CA Bundle." -category = "main" optional = false python-versions = ">=3.6" files = [ @@ -220,7 +208,6 @@ files = [ name = "charset-normalizer" version = "3.1.0" description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet." -category = "main" optional = false python-versions = ">=3.7.0" files = [ @@ -305,7 +292,6 @@ files = [ name = "click" version = "8.1.3" description = "Composable command line interface toolkit" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -320,7 +306,6 @@ colorama = {version = "*", markers = "platform_system == \"Windows\""} name = "codefind" version = "0.1.3" description = "Find code objects and their referents" -category = "main" optional = false python-versions = ">=3.8,<4.0" files = [ @@ -332,7 +317,6 @@ files = [ name = "coleo" version = "0.3.2" description = "The nicest way to develop a command-line interface" -category = "main" optional = false python-versions = ">=3.7,<4.0" files = [ @@ -351,7 +335,6 @@ yaml = ["pyyaml (>=5.3,<6.0)"] name = "colorama" version = "0.4.6" description = "Cross-platform colored terminal text." -category = "main" optional = false python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7" files = [ @@ -363,7 +346,6 @@ files = [ name = "colorlog" version = "6.7.0" description = "Add colours to the output of Python's logging module." -category = "main" optional = false python-versions = ">=3.6" files = [ @@ -381,7 +363,6 @@ development = ["black", "flake8", "mypy", "pytest", "types-colorama"] name = "coverage" version = "7.2.7" description = "Code coverage measurement for Python" -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -457,7 +438,6 @@ toml = ["tomli"] name = "cp-template" version = "0.3.0" description = "A tool to copy templated directories" -category = "main" optional = false python-versions = ">=3.7,<4.0" files = [ @@ -473,7 +453,6 @@ pystache = ">=0.6.0,<0.7.0" name = "distlib" version = "0.3.6" description = "Distribution utilities" -category = "main" optional = false python-versions = "*" files = [ @@ -485,7 +464,6 @@ files = [ name = "dnspython" version = "2.3.0" description = "DNS toolkit" -category = "main" optional = false python-versions = ">=3.7,<4.0" files = [ @@ -506,7 +484,6 @@ wmi = ["wmi (>=1.5.1,<2.0.0)"] name = "docutils" version = "0.17.1" description = "Docutils -- Python Documentation Utilities" -category = "dev" optional = false python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" files = [ @@ -518,7 +495,6 @@ files = [ name = "executing" version = "1.2.0" description = "Get the currently executing AST node of a frame, and other information" -category = "main" optional = false python-versions = "*" files = [ @@ -533,7 +509,6 @@ tests = ["asttokens", "littleutils", "pytest", "rich"] name = "filelock" version = "3.12.1" description = "A platform independent file lock." -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -549,7 +524,6 @@ testing = ["covdefaults (>=2.3)", "coverage (>=7.2.7)", "diff-cover (>=7.5)", "p name = "flake8" version = "4.0.1" description = "the modular source code checker: pep8 pyflakes and co" -category = "dev" optional = false python-versions = ">=3.6" files = [ @@ -566,7 +540,6 @@ pyflakes = ">=2.4.0,<2.5.0" name = "gitdb" version = "4.0.10" description = "Git Object Database" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -581,7 +554,6 @@ smmap = ">=3.0.1,<6" name = "gitpython" version = "3.1.31" description = "GitPython is a Python library used to interact with Git repositories" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -596,7 +568,6 @@ gitdb = ">=4.0.1,<5" name = "giving" version = "0.4.2" description = "Reactive logging" -category = "main" optional = false python-versions = ">=3.7,<4.0" files = [ @@ -613,7 +584,6 @@ varname = ">=0.10.0,<0.11.0" name = "greenlet" version = "2.0.2" description = "Lightweight in-process concurrent programming" -category = "main" optional = false python-versions = ">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*" files = [ @@ -622,6 +592,7 @@ files = [ {file = "greenlet-2.0.2-cp27-cp27m-win32.whl", hash = "sha256:6c3acb79b0bfd4fe733dff8bc62695283b57949ebcca05ae5c129eb606ff2d74"}, {file = "greenlet-2.0.2-cp27-cp27m-win_amd64.whl", hash = "sha256:283737e0da3f08bd637b5ad058507e578dd462db259f7f6e4c5c365ba4ee9343"}, {file = "greenlet-2.0.2-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:d27ec7509b9c18b6d73f2f5ede2622441de812e7b1a80bbd446cb0633bd3d5ae"}, + {file = "greenlet-2.0.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d967650d3f56af314b72df7089d96cda1083a7fc2da05b375d2bc48c82ab3f3c"}, {file = "greenlet-2.0.2-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:30bcf80dda7f15ac77ba5af2b961bdd9dbc77fd4ac6105cee85b0d0a5fcf74df"}, {file = "greenlet-2.0.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:26fbfce90728d82bc9e6c38ea4d038cba20b7faf8a0ca53a9c07b67318d46088"}, {file = "greenlet-2.0.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9190f09060ea4debddd24665d6804b995a9c122ef5917ab26e1566dcc712ceeb"}, @@ -630,6 +601,7 @@ files = [ {file = "greenlet-2.0.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:76ae285c8104046b3a7f06b42f29c7b73f77683df18c49ab5af7983994c2dd91"}, {file = "greenlet-2.0.2-cp310-cp310-win_amd64.whl", hash = "sha256:2d4686f195e32d36b4d7cf2d166857dbd0ee9f3d20ae349b6bf8afc8485b3645"}, {file = "greenlet-2.0.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:c4302695ad8027363e96311df24ee28978162cdcdd2006476c43970b384a244c"}, + {file = "greenlet-2.0.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:d4606a527e30548153be1a9f155f4e283d109ffba663a15856089fb55f933e47"}, {file = "greenlet-2.0.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c48f54ef8e05f04d6eff74b8233f6063cb1ed960243eacc474ee73a2ea8573ca"}, {file = "greenlet-2.0.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:a1846f1b999e78e13837c93c778dcfc3365902cfb8d1bdb7dd73ead37059f0d0"}, {file = "greenlet-2.0.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a06ad5312349fec0ab944664b01d26f8d1f05009566339ac6f63f56589bc1a2"}, @@ -659,6 +631,7 @@ files = [ {file = "greenlet-2.0.2-cp37-cp37m-win32.whl", hash = "sha256:3f6ea9bd35eb450837a3d80e77b517ea5bc56b4647f5502cd28de13675ee12f7"}, {file = "greenlet-2.0.2-cp37-cp37m-win_amd64.whl", hash = "sha256:7492e2b7bd7c9b9916388d9df23fa49d9b88ac0640db0a5b4ecc2b653bf451e3"}, {file = "greenlet-2.0.2-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:b864ba53912b6c3ab6bcb2beb19f19edd01a6bfcbdfe1f37ddd1778abfe75a30"}, + {file = "greenlet-2.0.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:1087300cf9700bbf455b1b97e24db18f2f77b55302a68272c56209d5587c12d1"}, {file = "greenlet-2.0.2-cp38-cp38-manylinux2010_x86_64.whl", hash = "sha256:ba2956617f1c42598a308a84c6cf021a90ff3862eddafd20c3333d50f0edb45b"}, {file = "greenlet-2.0.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fc3a569657468b6f3fb60587e48356fe512c1754ca05a564f11366ac9e306526"}, {file = "greenlet-2.0.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8eab883b3b2a38cc1e050819ef06a7e6344d4a990d24d45bc6f2cf959045a45b"}, @@ -667,6 +640,7 @@ files = [ {file = "greenlet-2.0.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b0ef99cdbe2b682b9ccbb964743a6aca37905fda5e0452e5ee239b1654d37f2a"}, {file = "greenlet-2.0.2-cp38-cp38-win32.whl", hash = "sha256:b80f600eddddce72320dbbc8e3784d16bd3fb7b517e82476d8da921f27d4b249"}, {file = "greenlet-2.0.2-cp38-cp38-win_amd64.whl", hash = "sha256:4d2e11331fc0c02b6e84b0d28ece3a36e0548ee1a1ce9ddde03752d9b79bba40"}, + {file = "greenlet-2.0.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8512a0c38cfd4e66a858ddd1b17705587900dd760c6003998e9472b77b56d417"}, {file = "greenlet-2.0.2-cp39-cp39-macosx_11_0_x86_64.whl", hash = "sha256:88d9ab96491d38a5ab7c56dd7a3cc37d83336ecc564e4e8816dbed12e5aaefc8"}, {file = "greenlet-2.0.2-cp39-cp39-manylinux2010_x86_64.whl", hash = "sha256:561091a7be172ab497a3527602d467e2b3fbe75f9e783d8b8ce403fa414f71a6"}, {file = "greenlet-2.0.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971ce5e14dc5e73715755d0ca2975ac88cfdaefcaab078a284fea6cfabf866df"}, @@ -687,7 +661,6 @@ test = ["objgraph", "psutil"] name = "hrepr" version = "0.4.1" description = "Extensible HTML representation for Python objects." -category = "main" optional = false python-versions = ">=3.6,<4.0" files = [ @@ -702,7 +675,6 @@ ovld = ">=0.3.2,<0.4.0" name = "idna" version = "3.4" description = "Internationalized Domain Names in Applications (IDNA)" -category = "main" optional = false python-versions = ">=3.5" files = [ @@ -714,7 +686,6 @@ files = [ name = "imagesize" version = "1.4.1" description = "Getting image size from png/jpeg/jpeg2000/gif file" -category = "dev" optional = false python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" files = [ @@ -726,7 +697,6 @@ files = [ name = "importlib-metadata" version = "6.6.0" description = "Read metadata from Python packages" -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -742,11 +712,28 @@ docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker perf = ["ipython"] testing = ["flake8 (<5)", "flufl.flake8", "importlib-resources (>=1.3)", "packaging", "pyfakefs", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)", "pytest-perf (>=0.9.2)"] +[[package]] +name = "importlib-resources" +version = "6.1.0" +description = "Read resources from Python packages" +optional = false +python-versions = ">=3.8" +files = [ + {file = "importlib_resources-6.1.0-py3-none-any.whl", hash = "sha256:aa50258bbfa56d4e33fbd8aa3ef48ded10d1735f11532b8df95388cc6bdb7e83"}, + {file = "importlib_resources-6.1.0.tar.gz", hash = "sha256:9d48dcccc213325e810fd723e7fbb45ccb39f6cf5c31f00cf2b965f5f10f3cb9"}, +] + +[package.dependencies] +zipp = {version = ">=3.1.0", markers = "python_version < \"3.10\""} + +[package.extras] +docs = ["furo", "jaraco.packaging (>=9.3)", "jaraco.tidelift (>=1.4)", "rst.linker (>=1.9)", "sphinx (<7.2.5)", "sphinx (>=3.5)", "sphinx-lint"] +testing = ["pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=2.2)", "pytest-mypy (>=0.9.1)", "pytest-ruff", "zipp (>=3.17)"] + [[package]] name = "iniconfig" version = "2.0.0" description = "brain-dead simple config-ini parsing" -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -758,7 +745,6 @@ files = [ name = "isort" version = "5.12.0" description = "A Python utility / library to sort Python imports." -category = "dev" optional = false python-versions = ">=3.8.0" files = [ @@ -776,7 +762,6 @@ requirements-deprecated-finder = ["pip-api", "pipreqs"] name = "jinja2" version = "3.1.2" description = "A very fast and expressive template engine." -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -794,7 +779,6 @@ i18n = ["Babel (>=2.7)"] name = "jinxed" version = "1.2.0" description = "Jinxed Terminal Library" -category = "main" optional = false python-versions = "*" files = [ @@ -809,7 +793,6 @@ ansicon = {version = "*", markers = "platform_system == \"Windows\""} name = "markdown-it-py" version = "3.0.0" description = "Python port of markdown-it. Markdown parsing, done right!" -category = "main" optional = false python-versions = ">=3.8" files = [ @@ -834,7 +817,6 @@ testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"] name = "markupsafe" version = "2.1.3" description = "Safely add untrusted strings to HTML/XML markup." -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -858,6 +840,16 @@ files = [ {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:5bbe06f8eeafd38e5d0a4894ffec89378b6c6a625ff57e3028921f8ff59318ac"}, {file = "MarkupSafe-2.1.3-cp311-cp311-win32.whl", hash = "sha256:dd15ff04ffd7e05ffcb7fe79f1b98041b8ea30ae9234aed2a9168b5797c3effb"}, {file = "MarkupSafe-2.1.3-cp311-cp311-win_amd64.whl", hash = "sha256:134da1eca9ec0ae528110ccc9e48041e0828d79f24121a1a146161103c76e686"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-macosx_10_9_universal2.whl", hash = "sha256:f698de3fd0c4e6972b92290a45bd9b1536bffe8c6759c62471efaa8acb4c37bc"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:aa57bd9cf8ae831a362185ee444e15a93ecb2e344c8e52e4d721ea3ab6ef1823"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ffcc3f7c66b5f5b7931a5aa68fc9cecc51e685ef90282f4a82f0f5e9b704ad11"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:47d4f1c5f80fc62fdd7777d0d40a2e9dda0a05883ab11374334f6c4de38adffd"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1f67c7038d560d92149c060157d623c542173016c4babc0c1913cca0564b9939"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:9aad3c1755095ce347e26488214ef77e0485a3c34a50c5a5e2471dff60b9dd9c"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:14ff806850827afd6b07a5f32bd917fb7f45b046ba40c57abdb636674a8b559c"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:8f9293864fe09b8149f0cc42ce56e3f0e54de883a9de90cd427f191c346eb2e1"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-win32.whl", hash = "sha256:715d3562f79d540f251b99ebd6d8baa547118974341db04f5ad06d5ea3eb8007"}, + {file = "MarkupSafe-2.1.3-cp312-cp312-win_amd64.whl", hash = "sha256:1b8dd8c3fd14349433c79fa8abeb573a55fc0fdd769133baac1f5e07abf54aeb"}, {file = "MarkupSafe-2.1.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8e254ae696c88d98da6555f5ace2279cf7cd5b3f52be2b5cf97feafe883b58d2"}, {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cb0932dc158471523c9637e807d9bfb93e06a95cbf010f1a38b98623b929ef2b"}, {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9402b03f1a1b4dc4c19845e5c749e3ab82d5078d16a2a4c2cd2df62d57bb0707"}, @@ -894,7 +886,6 @@ files = [ name = "mccabe" version = "0.6.1" description = "McCabe checker, plugin for flake8" -category = "dev" optional = false python-versions = "*" files = [ @@ -906,7 +897,6 @@ files = [ name = "mdurl" version = "0.1.2" description = "Markdown URL utilities" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -918,7 +908,6 @@ files = [ name = "mypy-extensions" version = "1.0.0" description = "Type system extensions for programs checked with the mypy type checker." -category = "dev" optional = false python-versions = ">=3.5" files = [ @@ -930,7 +919,6 @@ files = [ name = "nox" version = "2021.10.1" description = "Flexible test automation." -category = "main" optional = false python-versions = ">=3.6" files = [ @@ -952,7 +940,6 @@ tox-to-nox = ["jinja2", "tox"] name = "numpy" version = "1.24.3" description = "Fundamental package for array computing in Python" -category = "main" optional = false python-versions = ">=3.8" files = [ @@ -990,7 +977,6 @@ files = [ name = "omegaconf" version = "2.3.0" description = "A flexible configuration library" -category = "main" optional = false python-versions = ">=3.6" files = [ @@ -999,14 +985,13 @@ files = [ ] [package.dependencies] -antlr4-python3-runtime = ">=4.9.0,<4.10.0" +antlr4-python3-runtime = "==4.9.*" PyYAML = ">=5.1.0" [[package]] name = "ovld" version = "0.3.2" description = "Overloading Python functions" -category = "main" optional = false python-versions = ">=3.6,<4.0" files = [ @@ -1018,7 +1003,6 @@ files = [ name = "packaging" version = "23.1" description = "Core utilities for Python packages" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1030,7 +1014,6 @@ files = [ name = "pandas" version = "1.5.3" description = "Powerful data structures for data analysis, time series, and statistics" -category = "main" optional = false python-versions = ">=3.8" files = [ @@ -1066,7 +1049,8 @@ files = [ [package.dependencies] numpy = [ {version = ">=1.20.3", markers = "python_version < \"3.10\""}, - {version = ">=1.21.0", markers = "python_version >= \"3.10\""}, + {version = ">=1.23.2", markers = "python_version >= \"3.11\""}, + {version = ">=1.21.0", markers = "python_version >= \"3.10\" and python_version < \"3.11\""}, ] python-dateutil = ">=2.8.1" pytz = ">=2020.1" @@ -1078,7 +1062,6 @@ test = ["hypothesis (>=5.5.3)", "pytest (>=6.0)", "pytest-xdist (>=1.31)"] name = "pathspec" version = "0.9.0" description = "Utility library for gitignore style pattern matching of file paths." -category = "main" optional = false python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,>=2.7" files = [ @@ -1090,7 +1073,6 @@ files = [ name = "pip" version = "23.1.2" description = "The PyPA recommended tool for installing Python packages." -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1102,7 +1084,6 @@ files = [ name = "pip-tools" version = "6.13.0" description = "pip-tools keeps your pinned dependencies fresh." -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1125,7 +1106,6 @@ testing = ["flit-core (>=2,<4)", "poetry-core (>=1.0.0)", "pytest (>=7.2.0)", "p name = "platformdirs" version = "3.5.3" description = "A small Python package for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1141,7 +1121,6 @@ test = ["appdirs (==1.4.4)", "covdefaults (>=2.3)", "pytest (>=7.3.1)", "pytest- name = "pluggy" version = "1.0.0" description = "plugin and hook calling mechanisms for python" -category = "dev" optional = false python-versions = ">=3.6" files = [ @@ -1157,7 +1136,6 @@ testing = ["pytest", "pytest-benchmark"] name = "psutil" version = "5.9.5" description = "Cross-platform lib for process and system monitoring in Python." -category = "main" optional = false python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" files = [ @@ -1184,7 +1162,6 @@ test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] name = "psycopg2-binary" version = "2.9.6" description = "psycopg2 - Python-PostgreSQL Database Adapter" -category = "main" optional = true python-versions = ">=3.6" files = [ @@ -1256,7 +1233,6 @@ files = [ name = "ptera" version = "1.4.1" description = "Call graph addressing library." -category = "main" optional = false python-versions = ">=3.7,<4.0" files = [ @@ -1272,7 +1248,6 @@ giving = ">=0.4.1,<0.5.0" name = "py" version = "1.11.0" description = "library with cross-python path, ini-parsing, io, code, log facilities" -category = "main" optional = false python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" files = [ @@ -1284,7 +1259,6 @@ files = [ name = "py-cpuinfo" version = "9.0.0" description = "Get CPU info with pure Python" -category = "main" optional = false python-versions = "*" files = [ @@ -1296,7 +1270,6 @@ files = [ name = "pycodestyle" version = "2.8.0" description = "Python style guide checker" -category = "dev" optional = false python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" files = [ @@ -1308,7 +1281,6 @@ files = [ name = "pyflakes" version = "2.4.0" description = "passive checker of Python programs" -category = "dev" optional = false python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" files = [ @@ -1320,7 +1292,6 @@ files = [ name = "pygments" version = "2.15.1" description = "Pygments is a syntax highlighting package written in Python." -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1335,7 +1306,6 @@ plugins = ["importlib-metadata"] name = "pymongo" version = "4.3.3" description = "Python driver for MongoDB " -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1430,7 +1400,6 @@ zstd = ["zstandard"] name = "pynvml" version = "11.5.0" description = "Python Bindings for the NVIDIA Management Library" -category = "main" optional = false python-versions = ">=3.6" files = [ @@ -1442,7 +1411,6 @@ files = [ name = "pyproject-hooks" version = "1.0.0" description = "Wrappers to call pyproject.toml-based build backend hooks." -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1457,7 +1425,6 @@ tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""} name = "pystache" version = "0.6.0" description = "Mustache for Python" -category = "main" optional = false python-versions = ">=3.6" files = [ @@ -1472,7 +1439,6 @@ test = ["nose"] name = "pytest" version = "6.2.5" description = "pytest: simple powerful testing with Python" -category = "dev" optional = false python-versions = ">=3.6" files = [ @@ -1497,7 +1463,6 @@ testing = ["argcomplete", "hypothesis (>=3.56)", "mock", "nose", "requests", "xm name = "pytest-cov" version = "3.0.0" description = "Pytest plugin for measuring coverage." -category = "dev" optional = false python-versions = ">=3.6" files = [ @@ -1516,7 +1481,6 @@ testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtuale name = "pytest-datadir" version = "1.4.1" description = "pytest plugin for test data directories and files" -category = "dev" optional = false python-versions = ">=3.6" files = [ @@ -1531,7 +1495,6 @@ pytest = ">=5.0" name = "pytest-regressions" version = "2.4.2" description = "Easy to use fixtures to write regression tests." -category = "dev" optional = false python-versions = ">=3.6" files = [ @@ -1554,7 +1517,6 @@ num = ["numpy", "pandas"] name = "python-dateutil" version = "2.8.2" description = "Extensions to the standard Python datetime module" -category = "main" optional = false python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" files = [ @@ -1569,7 +1531,6 @@ six = ">=1.5" name = "pytz" version = "2023.3" description = "World timezone definitions, modern and historical" -category = "main" optional = false python-versions = "*" files = [ @@ -1581,7 +1542,6 @@ files = [ name = "pyyaml" version = "6.0" description = "YAML parser and emitter for Python" -category = "main" optional = false python-versions = ">=3.6" files = [ @@ -1631,7 +1591,6 @@ files = [ name = "reactivex" version = "4.0.4" description = "ReactiveX (Rx) for Python" -category = "main" optional = false python-versions = ">=3.7,<4.0" files = [ @@ -1646,7 +1605,6 @@ typing-extensions = ">=4.1.1,<5.0.0" name = "requests" version = "2.31.0" description = "Python HTTP for Humans." -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1668,7 +1626,6 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] name = "rich" version = "13.4.2" description = "Render rich text, tables, progress bars, syntax highlighting, markdown and more to the terminal" -category = "main" optional = false python-versions = ">=3.7.0" files = [ @@ -1688,7 +1645,6 @@ jupyter = ["ipywidgets (>=7.5.1,<9)"] name = "setuptools" version = "67.8.0" description = "Easily download, build, install, upgrade, and uninstall Python packages" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1705,7 +1661,6 @@ testing-integration = ["build[virtualenv]", "filelock (>=3.4.0)", "jaraco.envs ( name = "six" version = "1.16.0" description = "Python 2 and 3 compatibility utilities" -category = "main" optional = false python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*" files = [ @@ -1717,7 +1672,6 @@ files = [ name = "smmap" version = "5.0.0" description = "A pure Python implementation of a sliding window memory map manager" -category = "main" optional = false python-versions = ">=3.6" files = [ @@ -1729,7 +1683,6 @@ files = [ name = "snowballstemmer" version = "2.2.0" description = "This package provides 29 stemmers for 28 languages generated from Snowball algorithms." -category = "dev" optional = false python-versions = "*" files = [ @@ -1741,7 +1694,6 @@ files = [ name = "sphinx" version = "4.5.0" description = "Python documentation generator" -category = "dev" optional = false python-versions = ">=3.6" files = [ @@ -1777,7 +1729,6 @@ test = ["cython", "html5lib", "pytest", "pytest-cov", "typed-ast"] name = "sphinx-rtd-theme" version = "1.2.2" description = "Read the Docs theme for Sphinx" -category = "dev" optional = false python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,>=2.7" files = [ @@ -1797,7 +1748,6 @@ dev = ["bump2version", "sphinxcontrib-httpdomain", "transifex-client", "wheel"] name = "sphinxcontrib-applehelp" version = "1.0.4" description = "sphinxcontrib-applehelp is a Sphinx extension which outputs Apple help books" -category = "dev" optional = false python-versions = ">=3.8" files = [ @@ -1813,7 +1763,6 @@ test = ["pytest"] name = "sphinxcontrib-devhelp" version = "1.0.2" description = "sphinxcontrib-devhelp is a sphinx extension which outputs Devhelp document." -category = "dev" optional = false python-versions = ">=3.5" files = [ @@ -1829,7 +1778,6 @@ test = ["pytest"] name = "sphinxcontrib-htmlhelp" version = "2.0.1" description = "sphinxcontrib-htmlhelp is a sphinx extension which renders HTML help files" -category = "dev" optional = false python-versions = ">=3.8" files = [ @@ -1845,7 +1793,6 @@ test = ["html5lib", "pytest"] name = "sphinxcontrib-jquery" version = "4.1" description = "Extension to include jQuery on newer Sphinx releases" -category = "dev" optional = false python-versions = ">=2.7" files = [ @@ -1860,7 +1807,6 @@ Sphinx = ">=1.8" name = "sphinxcontrib-jsmath" version = "1.0.1" description = "A sphinx extension which renders display math in HTML via JavaScript" -category = "dev" optional = false python-versions = ">=3.5" files = [ @@ -1875,7 +1821,6 @@ test = ["flake8", "mypy", "pytest"] name = "sphinxcontrib-qthelp" version = "1.0.3" description = "sphinxcontrib-qthelp is a sphinx extension which outputs QtHelp document." -category = "dev" optional = false python-versions = ">=3.5" files = [ @@ -1891,7 +1836,6 @@ test = ["pytest"] name = "sphinxcontrib-serializinghtml" version = "1.1.5" description = "sphinxcontrib-serializinghtml is a sphinx extension which outputs \"serialized\" HTML files (json and pickle)." -category = "dev" optional = false python-versions = ">=3.5" files = [ @@ -1907,7 +1851,6 @@ test = ["pytest"] name = "sqlalchemy" version = "2.0.16" description = "Database Abstraction Library" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -1986,7 +1929,6 @@ sqlcipher = ["sqlcipher3-binary"] name = "toml" version = "0.10.2" description = "Python Library for Tom's Obvious, Minimal Language" -category = "dev" optional = false python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*" files = [ @@ -1998,7 +1940,6 @@ files = [ name = "tomli" version = "2.0.1" description = "A lil' TOML parser" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -2010,7 +1951,6 @@ files = [ name = "tqdm" version = "4.65.0" description = "Fast, Extensible Progress Meter" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -2031,7 +1971,6 @@ telegram = ["requests"] name = "typing-extensions" version = "4.6.3" description = "Backported and Experimental Type Hints for Python 3.7+" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -2043,7 +1982,6 @@ files = [ name = "urllib3" version = "2.0.3" description = "HTTP library with thread-safe connection pooling, file post, and more." -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -2061,7 +1999,6 @@ zstd = ["zstandard (>=0.18.0)"] name = "varname" version = "0.10.0" description = "Dark magics about variable names in python." -category = "main" optional = false python-versions = ">=3.6,<4.0" files = [ @@ -2079,7 +2016,6 @@ all = ["asttokens (>=2.0.0,<3.0.0)", "pure_eval (<1.0.0)"] name = "virtualenv" version = "20.23.0" description = "Virtual Python Environment builder" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -2100,7 +2036,6 @@ test = ["covdefaults (>=2.3)", "coverage (>=7.2.3)", "coverage-enable-subprocess name = "voir" version = "0.2.10" description = "Instrument, extend and visualize your programs" -category = "main" optional = false python-versions = ">=3.7,<4.0" files = [ @@ -2120,7 +2055,6 @@ rich = ">=13.3.2,<14.0.0" name = "wcwidth" version = "0.2.6" description = "Measures the displayed width of unicode strings in a terminal" -category = "main" optional = false python-versions = "*" files = [ @@ -2132,7 +2066,6 @@ files = [ name = "wheel" version = "0.40.0" description = "A built-package format for Python" -category = "main" optional = false python-versions = ">=3.7" files = [ @@ -2147,7 +2080,6 @@ test = ["pytest (>=6.0.0)"] name = "zipp" version = "3.15.0" description = "Backport of pathlib-compatible object wrapper for zip files" -category = "dev" optional = false python-versions = ">=3.7" files = [ @@ -2161,5 +2093,5 @@ testing = ["big-O", "flake8 (<5)", "jaraco.functools", "jaraco.itertools", "more [metadata] lock-version = "2.0" -python-versions = ">=3.8,<3.11" -content-hash = "2852dd9dc4b604714a06f452ee27c295e946ad86b9955ac133da9ca3b92ad1f7" +python-versions = ">=3.8,<4.0" +content-hash = "0407b1f9e231b83ca25d848e4c21033a7016d5825c31a86ce075479b4b419fa8" diff --git a/pyproject.toml b/pyproject.toml index 8dca0260f..f349e2dea 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -6,7 +6,8 @@ authors = ["Olivier Breuleux "] license = "MIT" [tool.poetry.dependencies] -python = ">=3.8,<3.11" +voir = {git = "https://github.com/breuleux/voir", branch = "master"} +python = ">=3.8,<4.0" giving = "^0.4.0" ptera = "^1.2.0" coleo = "^0.3.0" @@ -19,7 +20,6 @@ hrepr = "^0.4.0" blessed = "^1.19.1" pathspec = "^0.9.0" cp-template = "^0.3.0" -voir = "^0.2.10" pandas = "^1.4.2" numpy = ">=1.23.0" pynvml = "^11.4.1" @@ -32,6 +32,7 @@ pymongo = "^4.3.3" psycopg2-binary = {version = "^2.9.6", optional = true} py-cpuinfo = "^9.0.0" psutil = "^5.9.5" +importlib-resources = "^6.1.0" [tool.poetry.dev-dependencies]