From 93862b1d9fa984ce73860080d170593eb1724bcc Mon Sep 17 00:00:00 2001 From: Setepenre Date: Fri, 2 Jun 2023 11:17:49 -0400 Subject: [PATCH 1/9] Fix Whisper test by adding --num-workers to hf benchmark suite (#137) (#139) Co-authored-by: Olivier Breuleux --- benchmarks/huggingface/bench/__main__.py | 10 +++++++++- config/base.yaml | 1 + 2 files changed, 10 insertions(+), 1 deletion(-) diff --git a/benchmarks/huggingface/bench/__main__.py b/benchmarks/huggingface/bench/__main__.py index 3d364a3b4..4e46f4ffd 100644 --- a/benchmarks/huggingface/bench/__main__.py +++ b/benchmarks/huggingface/bench/__main__.py @@ -40,7 +40,9 @@ def __init__(self, args): repeat=100000, generators=generators[info.category](info), ) - self.loader = DataLoader(self.data, batch_size=args.batch_size) + self.loader = DataLoader( + self.data, batch_size=args.batch_size, num_workers=args.num_workers + ) self.amp_scaler = torch.cuda.amp.GradScaler(enabled=is_fp16_allowed(args)) if is_fp16_allowed(args): @@ -130,6 +132,12 @@ def parser(): default="fp32", help="Precision configuration", ) + parser.add_argument( + "--num-workers", + type=int, + default=8, + help="number of workers for data loading", + ) # parser.add_argument( # "--no-stdout", # action="store_true", diff --git a/config/base.yaml b/config/base.yaml index 5abbccee0..e0920af3f 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -25,6 +25,7 @@ _hf: install_group: torch argv: --precision: 'tf32-fp16' + --num-workers: 8 plan: method: per_gpu From 4d0395afbaa42b14a1334d0749fcbe3cd2922cdb Mon Sep 17 00:00:00 2001 From: Setepenre Date: Fri, 2 Jun 2023 12:09:51 -0400 Subject: [PATCH 2/9] Disable ROCm tests while we find a suitable machine replacement (#140) (#141) --- .github/workflows/tests.yml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.github/workflows/tests.yml b/.github/workflows/tests.yml index 1deb971ba..e21392c2d 100644 --- a/.github/workflows/tests.yml +++ b/.github/workflows/tests.yml @@ -27,8 +27,8 @@ jobs: include: - arch: cuda exclude : "no-cuda" - - arch: rocm - exclude : "no-rocm" + # - arch: rocm + # exclude : "no-rocm" runs-on: [self-hosted, "${{ matrix.arch }}"] From b1390c24be0c95372fff4a38dc459aac9764ef6e Mon Sep 17 00:00:00 2001 From: Setepenre Date: Mon, 29 Jan 2024 10:46:21 -0500 Subject: [PATCH 3/9] Ignore verification (#186) * Ignore verification * update voir --------- Co-authored-by: pierre.delaunay --- .gitignore | 5 + .no_report | 0 benchmarks/accelerate_opt/main.py | 4 +- .../accelerate_opt/requirements.cuda.txt | 2 +- .../accelerate_opt/requirements.rocm.txt | 2 +- benchmarks/dlrm/requirements.cuda.txt | 2 +- benchmarks/dlrm/requirements.rocm.txt | 2 +- benchmarks/huggingface/requirements.cuda.txt | 2 +- benchmarks/huggingface/requirements.rocm.txt | 2 +- .../stable_baselines3/requirements.cuda.txt | 2 +- .../stable_baselines3/requirements.rocm.txt | 2 +- benchmarks/stargan/requirements.cuda.txt | 2 +- benchmarks/stargan/requirements.rocm.txt | 2 +- benchmarks/super-slomo/requirements.cuda.txt | 2 +- benchmarks/super-slomo/requirements.rocm.txt | 2 +- benchmarks/timm/requirements.cuda.txt | 2 +- benchmarks/timm/requirements.rocm.txt | 2 +- benchmarks/torchvision/requirements.cuda.txt | 2 +- benchmarks/torchvision/requirements.rocm.txt | 2 +- run.sh | 16 +++ slurm.sh | 129 ++++++++++++++++++ 21 files changed, 169 insertions(+), 17 deletions(-) create mode 100644 .no_report create mode 100644 run.sh create mode 100644 slurm.sh diff --git a/.gitignore b/.gitignore index 7237a95a5..39ba18454 100644 --- a/.gitignore +++ b/.gitignore @@ -11,3 +11,8 @@ dist/ .envrc docs/_build .pin-constraints-* +test.out +output/ +workspace/ +.pin/tmp-* + diff --git a/.no_report b/.no_report new file mode 100644 index 000000000..e69de29bb diff --git a/benchmarks/accelerate_opt/main.py b/benchmarks/accelerate_opt/main.py index 15c60463b..aacc1420f 100644 --- a/benchmarks/accelerate_opt/main.py +++ b/benchmarks/accelerate_opt/main.py @@ -145,17 +145,19 @@ def mblog(data): validation_split_percentage = config["validation_split_percentage"] dataset_name = config["dataset_name"] dataset_config_name = config["dataset_config_name"] - raw_datasets = load_dataset(dataset_name, dataset_config_name) + raw_datasets = load_dataset(dataset_name, dataset_config_name, ignore_verifications=True) if "validation" not in raw_datasets.keys(): raw_datasets["validation"] = load_dataset( dataset_name, dataset_config_name, split=f"train[:{validation_split_percentage}%]", + ignore_verifications=True, ) raw_datasets["train"] = load_dataset( dataset_name, dataset_config_name, split=f"train[{validation_split_percentage}%:]", + ignore_verifications=True, ) model_name = config["model_name"] diff --git a/benchmarks/accelerate_opt/requirements.cuda.txt b/benchmarks/accelerate_opt/requirements.cuda.txt index 6461568e6..47f1707a2 100644 --- a/benchmarks/accelerate_opt/requirements.cuda.txt +++ b/benchmarks/accelerate_opt/requirements.cuda.txt @@ -315,7 +315,7 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/accelerate_opt/requirements.in xxhash==3.2.0 # via diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index 11255be92..246b3e79a 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -315,7 +315,7 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/accelerate_opt/requirements.in xxhash==3.2.0 # via diff --git a/benchmarks/dlrm/requirements.cuda.txt b/benchmarks/dlrm/requirements.cuda.txt index ab2fd9286..51dd839c1 100644 --- a/benchmarks/dlrm/requirements.cuda.txt +++ b/benchmarks/dlrm/requirements.cuda.txt @@ -336,7 +336,7 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/dlrm/requirements.in websocket-client==1.5.1 # via diff --git a/benchmarks/dlrm/requirements.rocm.txt b/benchmarks/dlrm/requirements.rocm.txt index 11cb3bfb1..dc6be8d87 100644 --- a/benchmarks/dlrm/requirements.rocm.txt +++ b/benchmarks/dlrm/requirements.rocm.txt @@ -336,7 +336,7 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/dlrm/requirements.in websocket-client==1.5.1 # via diff --git a/benchmarks/huggingface/requirements.cuda.txt b/benchmarks/huggingface/requirements.cuda.txt index 089a92e2d..33da5f949 100644 --- a/benchmarks/huggingface/requirements.cuda.txt +++ b/benchmarks/huggingface/requirements.cuda.txt @@ -175,5 +175,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.rocm.txt b/benchmarks/huggingface/requirements.rocm.txt index 98666feb6..aea7d619d 100644 --- a/benchmarks/huggingface/requirements.rocm.txt +++ b/benchmarks/huggingface/requirements.rocm.txt @@ -175,5 +175,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/stable_baselines3/requirements.cuda.txt b/benchmarks/stable_baselines3/requirements.cuda.txt index 14b999925..79594a401 100644 --- a/benchmarks/stable_baselines3/requirements.cuda.txt +++ b/benchmarks/stable_baselines3/requirements.cuda.txt @@ -304,7 +304,7 @@ urllib3==1.26.15 # sentry-sdk varname==0.10.0 # via giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/stable_baselines3/requirements.in wandb==0.14.0 # via -r benchmarks/stable_baselines3/requirements.in diff --git a/benchmarks/stable_baselines3/requirements.rocm.txt b/benchmarks/stable_baselines3/requirements.rocm.txt index 15439c8bd..70f4dfd5b 100644 --- a/benchmarks/stable_baselines3/requirements.rocm.txt +++ b/benchmarks/stable_baselines3/requirements.rocm.txt @@ -304,7 +304,7 @@ urllib3==1.26.15 # sentry-sdk varname==0.10.0 # via giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/stable_baselines3/requirements.in wandb==0.14.0 # via -r benchmarks/stable_baselines3/requirements.in diff --git a/benchmarks/stargan/requirements.cuda.txt b/benchmarks/stargan/requirements.cuda.txt index 6aa5d81d8..397099c14 100644 --- a/benchmarks/stargan/requirements.cuda.txt +++ b/benchmarks/stargan/requirements.cuda.txt @@ -152,5 +152,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/stargan/requirements.in diff --git a/benchmarks/stargan/requirements.rocm.txt b/benchmarks/stargan/requirements.rocm.txt index 2b05df289..d087a07fe 100644 --- a/benchmarks/stargan/requirements.rocm.txt +++ b/benchmarks/stargan/requirements.rocm.txt @@ -152,5 +152,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/stargan/requirements.in diff --git a/benchmarks/super-slomo/requirements.cuda.txt b/benchmarks/super-slomo/requirements.cuda.txt index efd7a7a41..f07ef8202 100644 --- a/benchmarks/super-slomo/requirements.cuda.txt +++ b/benchmarks/super-slomo/requirements.cuda.txt @@ -157,5 +157,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.rocm.txt b/benchmarks/super-slomo/requirements.rocm.txt index fc6ab3862..1e08a2ed2 100644 --- a/benchmarks/super-slomo/requirements.rocm.txt +++ b/benchmarks/super-slomo/requirements.rocm.txt @@ -157,5 +157,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/timm/requirements.cuda.txt b/benchmarks/timm/requirements.cuda.txt index 7c60c6953..872840cc3 100644 --- a/benchmarks/timm/requirements.cuda.txt +++ b/benchmarks/timm/requirements.cuda.txt @@ -168,5 +168,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.rocm.txt b/benchmarks/timm/requirements.rocm.txt index ffdf88437..33980fa4d 100644 --- a/benchmarks/timm/requirements.rocm.txt +++ b/benchmarks/timm/requirements.rocm.txt @@ -168,5 +168,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/timm/requirements.in diff --git a/benchmarks/torchvision/requirements.cuda.txt b/benchmarks/torchvision/requirements.cuda.txt index 6486a70c9..da06be073 100644 --- a/benchmarks/torchvision/requirements.cuda.txt +++ b/benchmarks/torchvision/requirements.cuda.txt @@ -154,5 +154,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.rocm.txt b/benchmarks/torchvision/requirements.rocm.txt index 1f44e9dad..639e07293 100644 --- a/benchmarks/torchvision/requirements.rocm.txt +++ b/benchmarks/torchvision/requirements.rocm.txt @@ -154,5 +154,5 @@ varname==0.10.0 # via # -c .pin-constraints-torch.txt # giving -voir==0.2.9 +voir==0.2.12 # via -r benchmarks/torchvision/requirements.in diff --git a/run.sh b/run.sh new file mode 100644 index 000000000..c01f2bfd3 --- /dev/null +++ b/run.sh @@ -0,0 +1,16 @@ + +OUTPUT="test.out" +rm -rf $OUTPUT +touch $OUTPUT +sbatch --ntasks=1\ + --gpus-per-task=rtx8000:1\ + --cpus-per-task=4\ + --time=01:30:00\ + --ntasks-per-node=1\ + --mem=64G\ + -o $OUTPUT\ + slurm.sh\ + -a cuda\ + -b stable_update + +tail -f $OUTPUT \ No newline at end of file diff --git a/slurm.sh b/slurm.sh new file mode 100644 index 000000000..abd3f9447 --- /dev/null +++ b/slurm.sh @@ -0,0 +1,129 @@ +#!/bin/bash + +function usage() { + echo "Usage: $0 [-m] [-p]" + echo " -h Display this help message." + echo " -a arch GPU arch (default: cuda)" + echo " -b BRANCH Branch to checkout (default: master)" + echo " -o ORIGIN Origin to use (default: github/mila/milabench)" + echo " -c CONFIG Configuration (default: milabench/config/standard.yaml)" + echo " -e ENV Environment (default: ./env)" + echo " -p PYTHON Python version (default: 3.9)" + echo " ARGUMENT Any additional argument you want to process." + exit 1 +} + +ARCH="cuda" +PYTHON="3.9" +BRANCH="master" +ORIGIN="https://github.com/mila-iqia/milabench.git" +LOC="$SLURM_TMPDIR" +CONFIG="$LOC/milabench/config/standard.yaml" +BASE="$LOC/base" +ENV="./env" +REMAINING_ARGS="" + + +while getopts ":hm:p:e:a:b:o:c:" opt; do + case $opt in + h) + usage + ;; + p) + PYTHON="$OPTARG" + ;; + b) + BRANCH="$OPTARG" + ;; + o) + ORIGIN="$OPTARG" + ;; + c) + CONFIG="$OPTARG" + ;; + e) + ENV="$OPTARG" + ;; + a) + ARCH="$OPTARG" + ;; + l) + # FIX ME + LOC="$OPTARG" + CONFIG="$LOC/milabench/config/standard.yaml" + BASE="$LOC/base" + ;; + :) + echo "Option -$OPTARG requires an argument." >&2 + usage + ;; + esac +done + +shift "$((OPTIND-1))" +REMAINING_ARGS="$@" + +echo " PYTHON: $PYTHON" +echo " branch: $BRANCH" +echo " origin: $ORIGIN" +echo " config: $CONFIG" +echo " env: $ENV" +echo " args: $REMAINING_ARGS" +# +# Fix problem with conda saying it is not "init properly" +# +CONDA_EXEC="$(which conda)" +CONDA_BASE=$(dirname $CONDA_EXEC) +source $CONDA_BASE/../etc/profile.d/conda.sh + +if [ -e $HOME/.credentials.env ]; then + source $HOME/.credentials.env +fi + +cd $LOC +# +# Create a new environment +# +if [ ! -d "$ENV" ] && [ "$ENV" != "base" ] && [ ! -d "$CONDA_ENVS/$ENV" ]; then + conda create --prefix $ENV python=$PYTHON -y +fi +conda activate $ENV + +export HF_HOME=$BASE/cache +export HF_DATASETS_CACHE=$BASE/cache +export TORCH_HOME=$BASE/cache +export XDG_CACHE_HOME=$BASE/cache +export MILABENCH_GPU_ARCH=$ARCH + +export MILABENCH_DASH=no +export PYTHONUNBUFFERED=1 +export MILABENCH_BASE=$BASE +export MILABENCH_CONFIG=$CONFIG +# +# Fetch the repo +# +git clone --single-branch --depth 1 -b $BRANCH $ORIGIN +python -m pip install -e ./milabench + +module load gcc/9.3.0 +module load cuda/11.8 + +echo "" +echo "Install" +echo "-------" +milabench install --config $CONFIG --base $BASE $REMAINING_ARGS + + +echo "" +echo "Prepare" +echo "-------" +milabench prepare --config $CONFIG --base $BASE $REMAINING_ARGS + +echo "" +echo "Run" +echo "---" +milabench run --config $CONFIG --base $BASE $REMAINING_ARGS + +echo "----" +echo "Done after $SECONDS" +echo "" From 2dd2b141d1974e26e042c26d3c57b753c73fffd0 Mon Sep 17 00:00:00 2001 From: Setepenre Date: Mon, 29 Jan 2024 15:27:31 -0500 Subject: [PATCH 4/9] Update main.py (#190) --- benchmarks/accelerate_opt/main.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/benchmarks/accelerate_opt/main.py b/benchmarks/accelerate_opt/main.py index aacc1420f..61bf1fa1b 100644 --- a/benchmarks/accelerate_opt/main.py +++ b/benchmarks/accelerate_opt/main.py @@ -145,19 +145,19 @@ def mblog(data): validation_split_percentage = config["validation_split_percentage"] dataset_name = config["dataset_name"] dataset_config_name = config["dataset_config_name"] - raw_datasets = load_dataset(dataset_name, dataset_config_name, ignore_verifications=True) + raw_datasets = load_dataset(dataset_name, dataset_config_name, revision="f5562967961a45407fa15044c5535a607200983f") if "validation" not in raw_datasets.keys(): raw_datasets["validation"] = load_dataset( dataset_name, dataset_config_name, split=f"train[:{validation_split_percentage}%]", - ignore_verifications=True, + revision="f5562967961a45407fa15044c5535a607200983f", ) raw_datasets["train"] = load_dataset( dataset_name, dataset_config_name, split=f"train[{validation_split_percentage}%:]", - ignore_verifications=True, + revision="f5562967961a45407fa15044c5535a607200983f", ) model_name = config["model_name"] From 42e41e5e678b725afcc96dcd55f146f4c46667e5 Mon Sep 17 00:00:00 2001 From: Setepenre Date: Tue, 6 Feb 2024 09:53:49 -0500 Subject: [PATCH 5/9] Instructions Example (#167) --- scripts/instructions.sh | 110 ++++++++++++++++++++++++++++++++++++++++ 1 file changed, 110 insertions(+) create mode 100644 scripts/instructions.sh diff --git a/scripts/instructions.sh b/scripts/instructions.sh new file mode 100644 index 000000000..e985babf0 --- /dev/null +++ b/scripts/instructions.sh @@ -0,0 +1,110 @@ +#!/bin/bash + + +set -m + +# +# +# + +echo ">> Configure the benchmark" +echo "==========================" + + +# +# Tweak the values to fit your system +# + +USERNAME=${USER:-"mila"} +SSH_KEY_FILE=$HOME/.ssh/id_rsa +ARCH="cuda" +WORKER_0="cn-d003" +WORKER_1="cn-d004" + + + +# Derived +VERSION="v0.0.8" +IMAGE="ghcr.io/mila-iqia/milabench:$ARCH-$VERSION" + + +# Create the config file +cat >overrides.yaml <> Prepare docker images" +echo "========================" + +ssh $USERNAME@$WORKER_0 "docker pull $IMAGE"& +ssh $USERNAME@$WORKER_1 "docker pull $IMAGE"& +fg +fg + +echo "<< =====================" +echo "" + +# +# +# + +echo ">> Run milabench" +echo "================" + +if [ "$ARCH" = "cuda" ]; then + docker run -it --rm --gpus all --network host --ipc=host --privileged \ + -v $SSH_KEY_FILE:/milabench/id_milabench \ + -v $(pwd)/results:/milabench/envs/runs \ + $IMAGE \ + milabench run --override "$(cat overrides.yaml)" + +elif [ "$ARCH" = "rocm" ]; then + docker run -it --rm --network host --ipc host --privileged \ + --security-opt seccomp=unconfined --group-add video \ + -v /opt/amdgpu/share/libdrm/amdgpu.ids:/opt/amdgpu/share/libdrm/amdgpu.ids \ + -v /opt/rocm:/opt/rocm \ + -v $(pwd)/results:/milabench/envs/runs \ + $IMAGE \ + milabench run --override "$(cat overrides.yaml)" +fi + +echo "<< =============" +echo "" + +# +# +# + +echo ">> Print report" +echo "===============" +docker run -it --rm \ + -v $(pwd)/results:/milabench/envs/runs \ + $IMAGE \ + milabench report --runs /milabench/envs/runs + +echo "<< ============" From b670cda606ff08deab1f0584fd66f739b52bbd0f Mon Sep 17 00:00:00 2001 From: Setepenre Date: Tue, 6 Feb 2024 09:55:54 -0500 Subject: [PATCH 6/9] ve scripts to folder (#194) --- run.sh => scripts/run.sh | 0 slurm.sh => scripts/slurm.sh | 0 2 files changed, 0 insertions(+), 0 deletions(-) rename run.sh => scripts/run.sh (100%) rename slurm.sh => scripts/slurm.sh (100%) diff --git a/run.sh b/scripts/run.sh similarity index 100% rename from run.sh rename to scripts/run.sh diff --git a/slurm.sh b/scripts/slurm.sh similarity index 100% rename from slurm.sh rename to scripts/slurm.sh From c6233206e316bcf0a467fbd6acff7e17eb001497 Mon Sep 17 00:00:00 2001 From: Setepenre Date: Thu, 8 Feb 2024 08:58:45 -0500 Subject: [PATCH 7/9] update pins (#191) Co-authored-by: Pierre Delaunay --- .gitignore | 7 +- .../accelerate_opt/requirements.cuda.txt | 194 +++++++++++----- .../accelerate_opt/requirements.rocm.txt | 208 ++++++++++++----- benchmarks/dlrm/requirements.cuda.txt | 203 ++++++++-------- benchmarks/dlrm/requirements.rocm.txt | 219 +++++++++--------- benchmarks/huggingface/requirements.cuda.txt | 120 +++++++--- benchmarks/huggingface/requirements.rocm.txt | 132 +++++++---- benchmarks/stargan/requirements.cuda.txt | 103 +++++--- benchmarks/stargan/requirements.rocm.txt | 115 ++++++--- benchmarks/super-slomo/requirements.cuda.txt | 107 ++++++--- benchmarks/super-slomo/requirements.rocm.txt | 119 +++++++--- benchmarks/timm/requirements.cuda.txt | 112 ++++++--- benchmarks/timm/requirements.rocm.txt | 124 +++++++--- benchmarks/torchvision/requirements.cuda.txt | 105 ++++++--- benchmarks/torchvision/requirements.rocm.txt | 117 +++++++--- 15 files changed, 1304 insertions(+), 681 deletions(-) diff --git a/.gitignore b/.gitignore index 39ba18454..b34d80aa2 100644 --- a/.gitignore +++ b/.gitignore @@ -11,8 +11,13 @@ dist/ .envrc docs/_build .pin-constraints-* + +scripts/inventory.yaml +output/ +sqlite.db +.ruff_cache/ + test.out output/ workspace/ .pin/tmp-* - diff --git a/benchmarks/accelerate_opt/requirements.cuda.txt b/benchmarks/accelerate_opt/requirements.cuda.txt index 47f1707a2..08ae318ea 100644 --- a/benchmarks/accelerate_opt/requirements.cuda.txt +++ b/benchmarks/accelerate_opt/requirements.cuda.txt @@ -2,13 +2,13 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=benchmarks/accelerate_opt/requirements.cu118.txt --resolver=backtracking .pin-constraints-accelerate.txt benchmarks/accelerate_opt/requirements.in +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.cuda.txt --resolver=backtracking .pin-constraints-opt.txt benchmarks/accelerate_opt/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 -accelerate==0.18.0 +accelerate==0.26.1 # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.4 +aiohttp==3.9.3 # via # -c .pin-constraints-torch.txt # datasets @@ -17,74 +17,86 @@ aiosignal==1.3.1 # via # -c .pin-constraints-torch.txt # aiohttp +annotated-types==0.6.0 + # via + # -c .pin-constraints-torch.txt + # pydantic antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -async-timeout==4.0.2 +async-timeout==4.0.3 # via # -c .pin-constraints-torch.txt # aiohttp -attrs==22.2.0 +asyncssh==2.14.2 + # via -r benchmarks/accelerate_opt/requirements.in +attrs==23.2.0 # via # -c .pin-constraints-torch.txt # aiohttp -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +cffi==1.16.0 # via # -c .pin-constraints-torch.txt - # aiohttp - # requests -cmake==3.26.1 + # cryptography +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt - # triton + # requests codefind==0.1.3 # via # -c .pin-constraints-torch.txt # ptera -datasets==2.11.0 +cryptography==42.0.2 + # via + # -c .pin-constraints-torch.txt + # asyncssh +datasets==2.16.1 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.8.3 +deepspeed==0.13.1 # via -r benchmarks/accelerate_opt/requirements.in -dill==0.3.6 +dill==0.3.7 # via # -c .pin-constraints-torch.txt # datasets # evaluate # multiprocess -evaluate==0.4.0 +evaluate==0.4.1 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 # via # -c .pin-constraints-torch.txt + # datasets # huggingface-hub # torch # transformers # triton -frozenlist==1.3.3 +frozenlist==1.4.1 # via # -c .pin-constraints-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.3.0 +fsspec[http]==2023.10.0 # via # -c .pin-constraints-torch.txt # datasets # evaluate + # huggingface-hub + # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt @@ -94,30 +106,28 @@ hjson==3.1.0 # via # -c .pin-constraints-torch.txt # deepspeed -huggingface-hub==0.13.3 +huggingface-hub==0.20.3 # via # -c .pin-constraints-torch.txt + # accelerate # datasets # evaluate + # tokenizers # transformers -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests # yarl -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -129,25 +139,25 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -multidict==6.0.4 +multidict==6.0.5 # via # -c .pin-constraints-torch.txt # aiohttp # yarl -multiprocess==0.70.14 +multiprocess==0.70.15 # via # -c .pin-constraints-torch.txt # datasets # evaluate -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -ninja==1.11.1 +ninja==1.11.1.1 # via # -c .pin-constraints-torch.txt # deepspeed -numpy==1.24.2 +numpy==1.26.3 # via # -c .pin-constraints-torch.txt # accelerate @@ -158,6 +168,52 @@ numpy==1.24.2 # pyarrow # torchvision # transformers +nvidia-cublas-cu11==11.11.3.6 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu11 + # nvidia-cusolver-cu11 + # torch +nvidia-cuda-cupti-cu11==11.8.87 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu11==8.7.0.84 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu11==10.9.0.58 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu11==10.3.0.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu11==11.4.1.48 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu11==11.7.5.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nccl-cu11==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvtx-cu11==11.8.86 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -166,7 +222,7 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -packaging==23.0 +packaging==23.2 # via # -c .pin-constraints-torch.txt # accelerate @@ -175,16 +231,16 @@ packaging==23.0 # evaluate # huggingface-hub # transformers -pandas==2.0.0 +pandas==2.2.0 # via # -c .pin-constraints-torch.txt # datasets # evaluate -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision -psutil==5.9.4 +psutil==5.9.8 # via # -c .pin-constraints-torch.txt # accelerate @@ -197,31 +253,44 @@ py-cpuinfo==9.0.0 # via # -c .pin-constraints-torch.txt # deepspeed -pyarrow==11.0.0 +pyarrow==15.0.0 + # via + # -c .pin-constraints-torch.txt + # datasets +pyarrow-hotfix==0.6 # via # -c .pin-constraints-torch.txt # datasets -pydantic==1.10.7 +pycparser==2.21 + # via + # -c .pin-constraints-torch.txt + # cffi +pydantic==2.6.0 # via # -c .pin-constraints-torch.txt # deepspeed -pygments==2.14.0 +pydantic-core==2.16.1 + # via + # -c .pin-constraints-torch.txt + # pydantic +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich pynvml==11.5.0 # via # -c .pin-constraints-torch.txt + # deepspeed # voir python-dateutil==2.8.2 # via # -c .pin-constraints-torch.txt # pandas -pytz==2023.3 +pytz==2024.1 # via # -c .pin-constraints-torch.txt # pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # accelerate @@ -233,11 +302,11 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -regex==2023.3.23 +regex==2023.12.25 # via # -c .pin-constraints-torch.txt # transformers -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # datasets @@ -250,38 +319,41 @@ requests==2.28.2 responses==0.18.0 # via # -c .pin-constraints-torch.txt - # datasets # evaluate -rich==13.3.3 +rich==13.7.0 # via # -r benchmarks/accelerate_opt/requirements.in # voir +safetensors==0.4.2 + # via + # -c .pin-constraints-torch.txt + # accelerate + # transformers six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens # python-dateutil -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -tokenizers==0.13.2 +tokenizers==0.15.1 # via # -c .pin-constraints-torch.txt # transformers -torch==2.0.0+cu118 +torch==2.2.0+cu118 # via # -r benchmarks/accelerate_opt/requirements.in # accelerate # deepspeed # torchaudio # torchvision - # triton -torchaudio==2.0.1+cu118 +torchaudio==2.2.0+cu118 # via -r benchmarks/accelerate_opt/requirements.in -torchvision==0.15.1+cu118 +torchvision==0.17.0+cu118 # via -r benchmarks/accelerate_opt/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin-constraints-torch.txt # datasets @@ -289,24 +361,26 @@ tqdm==4.65.0 # evaluate # huggingface-hub # transformers -transformers==4.27.4 +transformers==4.37.2 # via -r benchmarks/accelerate_opt/requirements.in -triton==2.0.0 +triton==2.2.0 # via # -c .pin-constraints-torch.txt # torch -typing-extensions==4.5.0 +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt + # asyncssh # huggingface-hub # pydantic + # pydantic-core # reactivex # torch -tzdata==2023.3 +tzdata==2023.4 # via # -c .pin-constraints-torch.txt # pandas -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests @@ -317,12 +391,12 @@ varname==0.10.0 # giving voir==0.2.12 # via -r benchmarks/accelerate_opt/requirements.in -xxhash==3.2.0 +xxhash==3.4.1 # via # -c .pin-constraints-torch.txt # datasets # evaluate -yarl==1.8.2 +yarl==1.9.4 # via # -c .pin-constraints-torch.txt # aiohttp diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index 246b3e79a..3d3ee113a 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -2,13 +2,13 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=benchmarks/accelerate_opt/requirements.rocm.txt --resolver=backtracking .pin-constraints-accelerate.txt benchmarks/accelerate_opt/requirements.in +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.rocm.txt --resolver=backtracking .pin-constraints-opt.txt benchmarks/accelerate_opt/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ -accelerate==0.18.0 +accelerate==0.26.1 # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.4 +aiohttp==3.9.3 # via # -c .pin-constraints-torch.txt # datasets @@ -17,74 +17,86 @@ aiosignal==1.3.1 # via # -c .pin-constraints-torch.txt # aiohttp +annotated-types==0.6.0 + # via + # -c .pin-constraints-torch.txt + # pydantic antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -async-timeout==4.0.2 +async-timeout==4.0.3 # via # -c .pin-constraints-torch.txt # aiohttp -attrs==22.2.0 +asyncssh==2.14.2 + # via -r benchmarks/accelerate_opt/requirements.in +attrs==23.2.0 # via # -c .pin-constraints-torch.txt # aiohttp -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +cffi==1.16.0 # via # -c .pin-constraints-torch.txt - # aiohttp - # requests -cmake==3.26.1 + # cryptography +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt - # pytorch-triton-rocm + # requests codefind==0.1.3 # via # -c .pin-constraints-torch.txt # ptera -datasets==2.11.0 +cryptography==42.0.2 + # via + # -c .pin-constraints-torch.txt + # asyncssh +datasets==2.16.1 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.8.3 +deepspeed==0.13.1 # via -r benchmarks/accelerate_opt/requirements.in -dill==0.3.6 +dill==0.3.7 # via # -c .pin-constraints-torch.txt # datasets # evaluate # multiprocess -evaluate==0.4.0 +evaluate==0.4.1 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 # via # -c .pin-constraints-torch.txt + # datasets # huggingface-hub - # pytorch-triton-rocm # torch # transformers -frozenlist==1.3.3 + # triton +frozenlist==1.4.1 # via # -c .pin-constraints-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.3.0 +fsspec[http]==2023.10.0 # via # -c .pin-constraints-torch.txt # datasets # evaluate + # huggingface-hub + # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt @@ -94,30 +106,28 @@ hjson==3.1.0 # via # -c .pin-constraints-torch.txt # deepspeed -huggingface-hub==0.13.3 +huggingface-hub==0.20.3 # via # -c .pin-constraints-torch.txt + # accelerate # datasets # evaluate + # tokenizers # transformers -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests # yarl -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -129,25 +139,25 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -multidict==6.0.4 +multidict==6.0.5 # via # -c .pin-constraints-torch.txt # aiohttp # yarl -multiprocess==0.70.14 +multiprocess==0.70.15 # via # -c .pin-constraints-torch.txt # datasets # evaluate -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -ninja==1.11.1 +ninja==1.11.1.1 # via # -c .pin-constraints-torch.txt # deepspeed -numpy==1.24.2 +numpy==1.26.3 # via # -c .pin-constraints-torch.txt # accelerate @@ -158,6 +168,58 @@ numpy==1.24.2 # pyarrow # torchvision # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -166,7 +228,7 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -packaging==23.0 +packaging==23.2 # via # -c .pin-constraints-torch.txt # accelerate @@ -175,16 +237,16 @@ packaging==23.0 # evaluate # huggingface-hub # transformers -pandas==2.0.0 +pandas==2.2.0 # via # -c .pin-constraints-torch.txt # datasets # evaluate -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision -psutil==5.9.4 +psutil==5.9.8 # via # -c .pin-constraints-torch.txt # accelerate @@ -197,35 +259,44 @@ py-cpuinfo==9.0.0 # via # -c .pin-constraints-torch.txt # deepspeed -pyarrow==11.0.0 +pyarrow==15.0.0 + # via + # -c .pin-constraints-torch.txt + # datasets +pyarrow-hotfix==0.6 # via # -c .pin-constraints-torch.txt # datasets -pydantic==1.10.7 +pycparser==2.21 + # via + # -c .pin-constraints-torch.txt + # cffi +pydantic==2.6.0 # via # -c .pin-constraints-torch.txt # deepspeed -pygments==2.14.0 +pydantic-core==2.16.1 + # via + # -c .pin-constraints-torch.txt + # pydantic +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich pynvml==11.5.0 # via # -c .pin-constraints-torch.txt + # deepspeed # voir python-dateutil==2.8.2 # via # -c .pin-constraints-torch.txt # pandas -pytorch-triton-rocm==2.0.1 - # via - # -c .pin-constraints-torch.txt - # torch -pytz==2023.3 +pytz==2024.1 # via # -c .pin-constraints-torch.txt # pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # accelerate @@ -237,11 +308,11 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -regex==2023.3.23 +regex==2023.12.25 # via # -c .pin-constraints-torch.txt # transformers -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # datasets @@ -254,38 +325,41 @@ requests==2.28.2 responses==0.18.0 # via # -c .pin-constraints-torch.txt - # datasets # evaluate -rich==13.3.3 +rich==13.7.0 # via # -r benchmarks/accelerate_opt/requirements.in # voir +safetensors==0.4.2 + # via + # -c .pin-constraints-torch.txt + # accelerate + # transformers six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens # python-dateutil -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -tokenizers==0.13.2 +tokenizers==0.15.1 # via # -c .pin-constraints-torch.txt # transformers -torch==2.0.0+rocm5.4.2 +torch==2.2.0 # via # -r benchmarks/accelerate_opt/requirements.in # accelerate # deepspeed - # pytorch-triton-rocm # torchaudio # torchvision -torchaudio==2.0.1+rocm5.4.2 +torchaudio==2.2.0 # via -r benchmarks/accelerate_opt/requirements.in -torchvision==0.15.1+rocm5.4.2 +torchvision==0.17.0 # via -r benchmarks/accelerate_opt/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin-constraints-torch.txt # datasets @@ -293,20 +367,26 @@ tqdm==4.65.0 # evaluate # huggingface-hub # transformers -transformers==4.27.4 +transformers==4.37.2 # via -r benchmarks/accelerate_opt/requirements.in -typing-extensions==4.5.0 +triton==2.2.0 + # via + # -c .pin-constraints-torch.txt + # torch +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt + # asyncssh # huggingface-hub # pydantic + # pydantic-core # reactivex # torch -tzdata==2023.3 +tzdata==2023.4 # via # -c .pin-constraints-torch.txt # pandas -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests @@ -317,12 +397,12 @@ varname==0.10.0 # giving voir==0.2.12 # via -r benchmarks/accelerate_opt/requirements.in -xxhash==3.2.0 +xxhash==3.4.1 # via # -c .pin-constraints-torch.txt # datasets # evaluate -yarl==1.8.2 +yarl==1.9.4 # via # -c .pin-constraints-torch.txt # aiohttp diff --git a/benchmarks/dlrm/requirements.cuda.txt b/benchmarks/dlrm/requirements.cuda.txt index 51dd839c1..708cb68aa 100644 --- a/benchmarks/dlrm/requirements.cuda.txt +++ b/benchmarks/dlrm/requirements.cuda.txt @@ -6,7 +6,7 @@ # --extra-index-url https://download.pytorch.org/whl/cu118 -absl-py==1.4.0 +absl-py==2.1.0 # via # -c .pin-constraints-torch.txt # tensorboard @@ -14,31 +14,27 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -cachetools==5.3.0 +cachetools==5.3.2 # via # -c .pin-constraints-torch.txt # google-auth -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # triton codefind==0.1.3 # via # -c .pin-constraints-torch.txt # ptera -docker==6.0.1 +docker==7.0.0 # via # -c .pin-constraints-torch.txt # torchx @@ -50,19 +46,20 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -fbgemm-gpu==0.4.1 +fbgemm-gpu==0.6.0+cu118 # via # -c .pin-constraints-torch.txt - # torchrec-nightly -filelock==3.10.7 + # torchrec +filelock==3.13.1 # via # -c .pin-constraints-torch.txt # torch # torchx # triton -fsspec==2023.3.0 +fsspec==2023.10.0 # via # -c .pin-constraints-torch.txt + # torch # torchx future==0.18.3 # via -r benchmarks/dlrm/requirements.in @@ -71,12 +68,12 @@ giving==0.4.2 # -c .pin-constraints-torch.txt # ptera # voir -google-auth==2.17.1 +google-auth==2.27.0 # via # -c .pin-constraints-torch.txt # google-auth-oauthlib # tensorboard -google-auth-oauthlib==1.0.0 +google-auth-oauthlib==1.2.0 # via # -c .pin-constraints-torch.txt # tensorboard @@ -84,44 +81,40 @@ graphviz==0.20.1 # via # -c .pin-constraints-torch.txt # torchviz -grpcio==1.53.0 +grpcio==1.60.1 # via # -c .pin-constraints-torch.txt # tensorboard -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -importlib-metadata==6.4.1 +importlib-metadata==7.0.1 # via # -c .pin-constraints-torch.txt # markdown # torchx -iopath==0.1.10 - # via - # -c .pin-constraints-torch.txt - # torchrec -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -joblib==1.2.0 +joblib==1.3.2 # via # -c .pin-constraints-torch.txt # scikit-learn -lit==16.0.0 +lightning-utilities==0.10.1 # via # -c .pin-constraints-torch.txt - # triton -markdown==3.4.3 + # torchmetrics +markdown==3.5.2 # via # -c .pin-constraints-torch.txt # tensorboard -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -138,19 +131,65 @@ mypy-extensions==1.0.0 # via # -c .pin-constraints-torch.txt # typing-inspect -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -r benchmarks/dlrm/requirements.in + # fbgemm-gpu # onnx - # pandas # scikit-learn # scipy # tensorboard # torchmetrics +nvidia-cublas-cu11==11.11.3.6 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu11 + # nvidia-cusolver-cu11 + # torch +nvidia-cuda-cupti-cu11==11.8.87 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu11==8.7.0.84 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu11==10.9.0.58 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu11==10.3.0.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu11==11.4.1.48 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu11==11.7.5.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nccl-cu11==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvtx-cu11==11.8.86 + # via + # -c .pin-constraints-torch.txt + # torch oauthlib==3.2.2 # via # -c .pin-constraints-torch.txt @@ -159,26 +198,19 @@ omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt # voir -onnx==1.13.1 +onnx==1.15.0 # via -r benchmarks/dlrm/requirements.in ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -packaging==23.0 +packaging==23.2 # via # -c .pin-constraints-torch.txt # docker + # lightning-utilities # torchmetrics -pandas==2.0.0 - # via - # -c .pin-constraints-torch.txt - # torchrec -portalocker==2.7.0 - # via - # -c .pin-constraints-torch.txt - # iopath -protobuf==3.20.3 +protobuf==4.23.4 # via # -c .pin-constraints-torch.txt # onnx @@ -187,18 +219,18 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pyasn1==0.4.8 +pyasn1==0.5.1 # via # -c .pin-constraints-torch.txt # pyasn1-modules # rsa -pyasn1-modules==0.2.8 +pyasn1-modules==0.3.0 # via # -c .pin-constraints-torch.txt # google-auth -pydot==1.4.2 +pydot==2.0.0 # via -r benchmarks/dlrm/requirements.in -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -206,7 +238,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pyparsing==3.0.9 +pyparsing==3.1.1 # via # -c .pin-constraints-torch.txt # pydot @@ -214,15 +246,7 @@ pyre-extensions==0.0.30 # via # -c .pin-constraints-torch.txt # torchx -python-dateutil==2.8.2 - # via - # -c .pin-constraints-torch.txt - # pandas -pytz==2023.3 - # via - # -c .pin-constraints-torch.txt - # pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # omegaconf @@ -231,7 +255,7 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # docker @@ -241,7 +265,7 @@ requests-oauthlib==1.3.1 # via # -c .pin-constraints-torch.txt # google-auth-oauthlib -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir @@ -249,9 +273,9 @@ rsa==4.9 # via # -c .pin-constraints-torch.txt # google-auth -scikit-learn==1.2.2 +scikit-learn==1.4.0 # via -r benchmarks/dlrm/requirements.in -scipy==1.10.1 +scipy==1.12.0 # via # -c .pin-constraints-torch.txt # scikit-learn @@ -259,74 +283,61 @@ six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens - # google-auth - # python-dateutil -sympy==1.11.1 + # tensorboard +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch tabulate==0.9.0 # via # -c .pin-constraints-torch.txt - # torchrec-nightly - # torchx-nightly -tensorboard==2.12.1 + # torchx +tensorboard==2.15.1 # via -r benchmarks/dlrm/requirements.in -tensorboard-data-server==0.7.0 +tensorboard-data-server==0.7.2 # via # -c .pin-constraints-torch.txt # tensorboard -tensorboard-plugin-wit==1.8.1 - # via - # -c .pin-constraints-torch.txt - # tensorboard -threadpoolctl==3.1.0 +threadpoolctl==3.2.0 # via # -c .pin-constraints-torch.txt # scikit-learn -torch==2.0.0+cu118 +torch==2.2.0+cu118 # via # -r benchmarks/dlrm/requirements.in # torchmetrics # torchviz - # triton -torchmetrics==0.11.4 +torchmetrics==1.0.3 # via # -c .pin-constraints-torch.txt # torchrec -torchrec==0.4.0 +torchrec==0.6.0+cu118 # via -r benchmarks/dlrm/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in torchx==0.5.0 # via -r benchmarks/dlrm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -r benchmarks/dlrm/requirements.in - # iopath # torchrec -triton==2.0.0 +triton==2.2.0 # via # -c .pin-constraints-torch.txt # torch -typing-extensions==4.5.0 +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt - # iopath - # onnx + # lightning-utilities # pyre-extensions # reactivex # torch # typing-inspect -typing-inspect==0.8.0 +typing-inspect==0.9.0 # via # -c .pin-constraints-torch.txt # pyre-extensions -tzdata==2023.3 - # via - # -c .pin-constraints-torch.txt - # pandas -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # docker @@ -338,19 +349,11 @@ varname==0.10.0 # giving voir==0.2.12 # via -r benchmarks/dlrm/requirements.in -websocket-client==1.5.1 - # via - # -c .pin-constraints-torch.txt - # docker -werkzeug==2.2.3 - # via - # -c .pin-constraints-torch.txt - # tensorboard -wheel==0.40.0 +werkzeug==3.0.1 # via # -c .pin-constraints-torch.txt # tensorboard -zipp==3.15.0 +zipp==3.17.0 # via # -c .pin-constraints-torch.txt # importlib-metadata diff --git a/benchmarks/dlrm/requirements.rocm.txt b/benchmarks/dlrm/requirements.rocm.txt index dc6be8d87..984a2eb79 100644 --- a/benchmarks/dlrm/requirements.rocm.txt +++ b/benchmarks/dlrm/requirements.rocm.txt @@ -6,7 +6,7 @@ # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ -absl-py==1.4.0 +absl-py==2.1.0 # via # -c .pin-constraints-torch.txt # tensorboard @@ -14,31 +14,27 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -cachetools==5.3.0 +cachetools==5.3.2 # via # -c .pin-constraints-torch.txt # google-auth -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm codefind==0.1.3 # via # -c .pin-constraints-torch.txt # ptera -docker==6.0.1 +docker==7.0.0 # via # -c .pin-constraints-torch.txt # torchx @@ -50,19 +46,20 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -fbgemm-gpu==0.4.1 +fbgemm-gpu==0.6.0 # via # -c .pin-constraints-torch.txt - # torchrec-nightly -filelock==3.10.7 + # torchrec +filelock==3.13.1 # via # -c .pin-constraints-torch.txt - # pytorch-triton-rocm # torch - # torchx-nightly -fsspec==2023.3.0 + # torchx + # triton +fsspec==2023.10.0 # via # -c .pin-constraints-torch.txt + # torch # torchx future==0.18.3 # via -r benchmarks/dlrm/requirements.in @@ -71,12 +68,12 @@ giving==0.4.2 # -c .pin-constraints-torch.txt # ptera # voir -google-auth==2.17.1 +google-auth==2.27.0 # via # -c .pin-constraints-torch.txt # google-auth-oauthlib # tensorboard -google-auth-oauthlib==1.0.0 +google-auth-oauthlib==1.2.0 # via # -c .pin-constraints-torch.txt # tensorboard @@ -84,44 +81,40 @@ graphviz==0.20.1 # via # -c .pin-constraints-torch.txt # torchviz -grpcio==1.53.0 +grpcio==1.60.1 # via # -c .pin-constraints-torch.txt # tensorboard -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -importlib-metadata==6.4.1 +importlib-metadata==7.0.1 # via # -c .pin-constraints-torch.txt # markdown # torchx -iopath==0.1.10 - # via - # -c .pin-constraints-torch.txt - # torchrec -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -joblib==1.2.0 +joblib==1.3.2 # via # -c .pin-constraints-torch.txt # scikit-learn -lit==16.0.0 +lightning-utilities==0.10.1 # via # -c .pin-constraints-torch.txt - # pytorch-triton-rocm -markdown==3.4.3 + # torchmetrics +markdown==3.5.2 # via # -c .pin-constraints-torch.txt # tensorboard -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -138,19 +131,71 @@ mypy-extensions==1.0.0 # via # -c .pin-constraints-torch.txt # typing-inspect -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -r benchmarks/dlrm/requirements.in + # fbgemm-gpu # onnx - # pandas # scikit-learn # scipy # tensorboard # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch oauthlib==3.2.2 # via # -c .pin-constraints-torch.txt @@ -159,26 +204,19 @@ omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt # voir -onnx==1.13.1 +onnx==1.15.0 # via -r benchmarks/dlrm/requirements.in ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -packaging==23.0 +packaging==23.2 # via # -c .pin-constraints-torch.txt # docker + # lightning-utilities # torchmetrics -pandas==2.0.0 - # via - # -c .pin-constraints-torch.txt - # torchrec -portalocker==2.7.0 - # via - # -c .pin-constraints-torch.txt - # iopath -protobuf==3.20.3 +protobuf==4.23.4 # via # -c .pin-constraints-torch.txt # onnx @@ -187,18 +225,18 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pyasn1==0.4.8 +pyasn1==0.5.1 # via # -c .pin-constraints-torch.txt # pyasn1-modules # rsa -pyasn1-modules==0.2.8 +pyasn1-modules==0.3.0 # via # -c .pin-constraints-torch.txt # google-auth -pydot==1.4.2 +pydot==2.0.0 # via -r benchmarks/dlrm/requirements.in -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -206,7 +244,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pyparsing==3.0.9 +pyparsing==3.1.1 # via # -c .pin-constraints-torch.txt # pydot @@ -214,19 +252,7 @@ pyre-extensions==0.0.30 # via # -c .pin-constraints-torch.txt # torchx -python-dateutil==2.8.2 - # via - # -c .pin-constraints-torch.txt - # pandas -pytorch-triton-rocm==2.0.1 - # via - # -c .pin-constraints-torch.txt - # torch -pytz==2023.3 - # via - # -c .pin-constraints-torch.txt - # pandas -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # omegaconf @@ -235,7 +261,7 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # docker @@ -245,7 +271,7 @@ requests-oauthlib==1.3.1 # via # -c .pin-constraints-torch.txt # google-auth-oauthlib -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir @@ -253,9 +279,9 @@ rsa==4.9 # via # -c .pin-constraints-torch.txt # google-auth -scikit-learn==1.2.2 +scikit-learn==1.4.0 # via -r benchmarks/dlrm/requirements.in -scipy==1.10.1 +scipy==1.12.0 # via # -c .pin-constraints-torch.txt # scikit-learn @@ -263,70 +289,61 @@ six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens - # google-auth - # python-dateutil -sympy==1.11.1 + # tensorboard +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch tabulate==0.9.0 # via # -c .pin-constraints-torch.txt - # torchrec-nightly - # torchx-nightly -tensorboard==2.12.1 + # torchx +tensorboard==2.15.1 # via -r benchmarks/dlrm/requirements.in -tensorboard-data-server==0.7.0 +tensorboard-data-server==0.7.2 # via # -c .pin-constraints-torch.txt # tensorboard -tensorboard-plugin-wit==1.8.1 - # via - # -c .pin-constraints-torch.txt - # tensorboard -threadpoolctl==3.1.0 +threadpoolctl==3.2.0 # via # -c .pin-constraints-torch.txt # scikit-learn -torch==2.0.0+rocm5.4.2 +torch==2.2.0 # via # -r benchmarks/dlrm/requirements.in - # pytorch-triton-rocm # torchmetrics # torchviz -torchmetrics==0.11.4 +torchmetrics==1.0.3 # via # -c .pin-constraints-torch.txt # torchrec -torchrec==0.4.0 +torchrec==0.6.0 # via -r benchmarks/dlrm/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in torchx==0.5.0 # via -r benchmarks/dlrm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -r benchmarks/dlrm/requirements.in - # iopath # torchrec -typing-extensions==4.5.0 +triton==2.2.0 # via # -c .pin-constraints-torch.txt - # iopath - # onnx + # torch +typing-extensions==4.9.0 + # via + # -c .pin-constraints-torch.txt + # lightning-utilities # pyre-extensions # reactivex # torch # typing-inspect -typing-inspect==0.8.0 +typing-inspect==0.9.0 # via # -c .pin-constraints-torch.txt # pyre-extensions -tzdata==2023.3 - # via - # -c .pin-constraints-torch.txt - # pandas -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # docker @@ -338,19 +355,11 @@ varname==0.10.0 # giving voir==0.2.12 # via -r benchmarks/dlrm/requirements.in -websocket-client==1.5.1 - # via - # -c .pin-constraints-torch.txt - # docker -werkzeug==2.2.3 - # via - # -c .pin-constraints-torch.txt - # tensorboard -wheel==0.40.0 +werkzeug==3.0.1 # via # -c .pin-constraints-torch.txt # tensorboard -zipp==3.15.0 +zipp==3.17.0 # via # -c .pin-constraints-torch.txt # importlib-metadata diff --git a/benchmarks/huggingface/requirements.cuda.txt b/benchmarks/huggingface/requirements.cuda.txt index 33da5f949..724105ae6 100644 --- a/benchmarks/huggingface/requirements.cuda.txt +++ b/benchmarks/huggingface/requirements.cuda.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=benchmarks/huggingface/requirements.cu118.txt --resolver=backtracking .pin-constraints-hf.txt benchmarks/huggingface/requirements.in +# pip-compile --output-file=benchmarks/huggingface/requirements.cuda.txt --resolver=backtracking .pin-constraints-hf.txt benchmarks/huggingface/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # triton codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,39 +30,41 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 # via # -c .pin-constraints-torch.txt # huggingface-hub # torch # transformers # triton +fsspec==2023.10.0 + # via + # -c .pin-constraints-torch.txt + # huggingface-hub + # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -huggingface-hub==0.13.3 +huggingface-hub==0.20.3 # via # -c .pin-constraints-torch.txt + # tokenizers # transformers -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -78,14 +76,60 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -c .pin-constraints-torch.txt # transformers +nvidia-cublas-cu11==11.11.3.6 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu11 + # nvidia-cusolver-cu11 + # torch +nvidia-cuda-cupti-cu11==11.8.87 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu11==8.7.0.84 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu11==10.9.0.58 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu11==10.3.0.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu11==11.4.1.48 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu11==11.7.5.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nccl-cu11==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvtx-cu11==11.8.86 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -94,7 +138,7 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -packaging==23.0 +packaging==23.2 # via # -c .pin-constraints-torch.txt # huggingface-hub @@ -103,7 +147,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -111,7 +155,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # huggingface-hub @@ -121,53 +165,55 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -regex==2023.3.23 +regex==2023.12.25 # via # -c .pin-constraints-torch.txt # transformers -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # huggingface-hub # transformers -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir +safetensors==0.4.2 + # via + # -c .pin-constraints-torch.txt + # transformers six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -tokenizers==0.13.2 +tokenizers==0.15.1 # via # -c .pin-constraints-torch.txt # transformers -torch==2.0.0+cu118 - # via - # -r benchmarks/huggingface/requirements.in - # triton -tqdm==4.65.0 +torch==2.2.0+cu118 + # via -r benchmarks/huggingface/requirements.in +tqdm==4.66.1 # via # -c .pin-constraints-torch.txt # huggingface-hub # transformers -transformers==4.27.4 +transformers==4.37.2 # via -r benchmarks/huggingface/requirements.in -triton==2.0.0 +triton==2.2.0 # via # -c .pin-constraints-torch.txt # torch -typing-extensions==4.5.0 +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # huggingface-hub # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/huggingface/requirements.rocm.txt b/benchmarks/huggingface/requirements.rocm.txt index aea7d619d..4c403925e 100644 --- a/benchmarks/huggingface/requirements.rocm.txt +++ b/benchmarks/huggingface/requirements.rocm.txt @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,39 +30,41 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 # via # -c .pin-constraints-torch.txt # huggingface-hub - # pytorch-triton-rocm # torch # transformers + # triton +fsspec==2023.10.0 + # via + # -c .pin-constraints-torch.txt + # huggingface-hub + # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -huggingface-hub==0.13.3 +huggingface-hub==0.20.3 # via # -c .pin-constraints-torch.txt + # tokenizers # transformers -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -78,14 +76,66 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -c .pin-constraints-torch.txt # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -94,7 +144,7 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -packaging==23.0 +packaging==23.2 # via # -c .pin-constraints-torch.txt # huggingface-hub @@ -103,7 +153,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -111,11 +161,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pytorch-triton-rocm==2.0.1 - # via - # -c .pin-constraints-torch.txt - # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # huggingface-hub @@ -125,49 +171,55 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -regex==2023.3.23 +regex==2023.12.25 # via # -c .pin-constraints-torch.txt # transformers -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # huggingface-hub # transformers -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir +safetensors==0.4.2 + # via + # -c .pin-constraints-torch.txt + # transformers six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -tokenizers==0.13.2 +tokenizers==0.15.1 # via # -c .pin-constraints-torch.txt # transformers -torch==2.0.0+rocm5.4.2 - # via - # -r benchmarks/huggingface/requirements.in - # pytorch-triton-rocm -tqdm==4.65.0 +torch==2.2.0 + # via -r benchmarks/huggingface/requirements.in +tqdm==4.66.1 # via # -c .pin-constraints-torch.txt # huggingface-hub # transformers -transformers==4.27.4 +transformers==4.37.2 # via -r benchmarks/huggingface/requirements.in -typing-extensions==4.5.0 +triton==2.2.0 + # via + # -c .pin-constraints-torch.txt + # torch +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # huggingface-hub # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/stargan/requirements.cuda.txt b/benchmarks/stargan/requirements.cuda.txt index 397099c14..5ab21901e 100644 --- a/benchmarks/stargan/requirements.cuda.txt +++ b/benchmarks/stargan/requirements.cuda.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=benchmarks/stargan/requirements.cu118.txt --resolver=backtracking .pin-constraints-stargan.txt benchmarks/stargan/requirements.in +# pip-compile --output-file=benchmarks/stargan/requirements.cuda.txt --resolver=backtracking .pin-constraints-stargan.txt benchmarks/stargan/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # triton codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,33 +30,33 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 # via # -c .pin-constraints-torch.txt # torch # triton +fsspec==2023.10.0 + # via + # -c .pin-constraints-torch.txt + # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -72,14 +68,60 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -r benchmarks/stargan/requirements.in # torchvision +nvidia-cublas-cu11==11.11.3.6 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu11 + # nvidia-cusolver-cu11 + # torch +nvidia-cuda-cupti-cu11==11.8.87 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu11==8.7.0.84 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu11==10.9.0.58 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu11==10.3.0.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu11==11.4.1.48 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu11==11.7.5.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nccl-cu11==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvtx-cu11==11.8.86 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -88,7 +130,7 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision @@ -96,7 +138,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -104,7 +146,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # omegaconf @@ -112,11 +154,11 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # torchvision -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir @@ -124,27 +166,26 @@ six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -torch==2.0.0+cu118 +torch==2.2.0+cu118 # via # -r benchmarks/stargan/requirements.in # torchvision - # triton -torchvision==0.15.1+cu118 +torchvision==0.17.0+cu118 # via -r benchmarks/stargan/requirements.in -triton==2.0.0 +triton==2.2.0 # via # -c .pin-constraints-torch.txt # torch -typing-extensions==4.5.0 +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/stargan/requirements.rocm.txt b/benchmarks/stargan/requirements.rocm.txt index d087a07fe..0cfc58c21 100644 --- a/benchmarks/stargan/requirements.rocm.txt +++ b/benchmarks/stargan/requirements.rocm.txt @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,33 +30,33 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 + # via + # -c .pin-constraints-torch.txt + # torch + # triton +fsspec==2023.10.0 # via # -c .pin-constraints-torch.txt - # pytorch-triton-rocm # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -72,14 +68,66 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -r benchmarks/stargan/requirements.in # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -88,7 +136,7 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision @@ -96,7 +144,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -104,11 +152,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pytorch-triton-rocm==2.0.1 - # via - # -c .pin-constraints-torch.txt - # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # omegaconf @@ -116,11 +160,11 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # torchvision -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir @@ -128,23 +172,26 @@ six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -torch==2.0.0+rocm5.4.2 +torch==2.2.0 # via # -r benchmarks/stargan/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.15.1+rocm5.4.2 +torchvision==0.17.0 # via -r benchmarks/stargan/requirements.in -typing-extensions==4.5.0 +triton==2.2.0 + # via + # -c .pin-constraints-torch.txt + # torch +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/super-slomo/requirements.cuda.txt b/benchmarks/super-slomo/requirements.cuda.txt index f07ef8202..06c2c257e 100644 --- a/benchmarks/super-slomo/requirements.cuda.txt +++ b/benchmarks/super-slomo/requirements.cuda.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=benchmarks/super-slomo/requirements.cu118.txt --resolver=backtracking .pin-constraints-super-slomo.txt benchmarks/super-slomo/requirements.in +# pip-compile --output-file=benchmarks/super-slomo/requirements.cuda.txt --resolver=backtracking .pin-constraints-super-slomo.txt benchmarks/super-slomo/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # triton codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,33 +30,33 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 # via # -c .pin-constraints-torch.txt # torch # triton +fsspec==2023.10.0 + # via + # -c .pin-constraints-torch.txt + # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -72,26 +68,72 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -r benchmarks/super-slomo/requirements.in # opencv-python # torchvision +nvidia-cublas-cu11==11.11.3.6 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu11 + # nvidia-cusolver-cu11 + # torch +nvidia-cuda-cupti-cu11==11.8.87 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu11==8.7.0.84 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu11==10.9.0.58 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu11==10.3.0.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu11==11.4.1.48 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu11==11.7.5.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nccl-cu11==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvtx-cu11==11.8.86 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt # voir -opencv-python==4.7.0.72 +opencv-python==4.9.0.80 # via -r benchmarks/super-slomo/requirements.in ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision @@ -99,7 +141,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -107,7 +149,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # omegaconf @@ -115,11 +157,11 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # torchvision -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir @@ -127,29 +169,28 @@ six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -torch==2.0.0+cu118 +torch==2.2.0+cu118 # via # -r benchmarks/super-slomo/requirements.in # torchvision - # triton -torchvision==0.15.1+cu118 +torchvision==0.17.0+cu118 # via -r benchmarks/super-slomo/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via -r benchmarks/super-slomo/requirements.in -triton==2.0.0 +triton==2.2.0 # via # -c .pin-constraints-torch.txt # torch -typing-extensions==4.5.0 +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/super-slomo/requirements.rocm.txt b/benchmarks/super-slomo/requirements.rocm.txt index 1e08a2ed2..7a2c83c28 100644 --- a/benchmarks/super-slomo/requirements.rocm.txt +++ b/benchmarks/super-slomo/requirements.rocm.txt @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,33 +30,33 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 + # via + # -c .pin-constraints-torch.txt + # torch + # triton +fsspec==2023.10.0 # via # -c .pin-constraints-torch.txt - # pytorch-triton-rocm # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -72,26 +68,78 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -r benchmarks/super-slomo/requirements.in # opencv-python # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt # voir -opencv-python==4.7.0.72 +opencv-python==4.9.0.80 # via -r benchmarks/super-slomo/requirements.in ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision @@ -99,7 +147,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -107,11 +155,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pytorch-triton-rocm==2.0.1 - # via - # -c .pin-constraints-torch.txt - # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # omegaconf @@ -119,11 +163,11 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # torchvision -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir @@ -131,25 +175,28 @@ six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -torch==2.0.0+rocm5.4.2 +torch==2.2.0 # via # -r benchmarks/super-slomo/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.15.1+rocm5.4.2 +torchvision==0.17.0 # via -r benchmarks/super-slomo/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via -r benchmarks/super-slomo/requirements.in -typing-extensions==4.5.0 +triton==2.2.0 + # via + # -c .pin-constraints-torch.txt + # torch +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/timm/requirements.cuda.txt b/benchmarks/timm/requirements.cuda.txt index 872840cc3..d320e9d41 100644 --- a/benchmarks/timm/requirements.cuda.txt +++ b/benchmarks/timm/requirements.cuda.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=benchmarks/timm/requirements.cu118.txt --resolver=backtracking .pin-constraints-timm.txt benchmarks/timm/requirements.in +# pip-compile --output-file=benchmarks/timm/requirements.cuda.txt --resolver=backtracking .pin-constraints-timm.txt benchmarks/timm/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # triton codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,36 +30,37 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 # via # -c .pin-constraints-torch.txt # huggingface-hub # torch # triton +fsspec==2023.10.0 + # via + # -c .pin-constraints-torch.txt + # huggingface-hub + # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -huggingface-hub==0.13.3 +huggingface-hub==0.20.3 # via -r benchmarks/timm/requirements.in -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -75,14 +72,60 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -c .pin-constraints-torch.txt # torchvision +nvidia-cublas-cu11==11.11.3.6 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu11 + # nvidia-cusolver-cu11 + # torch +nvidia-cuda-cupti-cu11==11.8.87 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu11==8.7.0.84 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu11==10.9.0.58 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu11==10.3.0.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu11==11.4.1.48 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu11==11.7.5.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nccl-cu11==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvtx-cu11==11.8.86 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -91,11 +134,11 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -packaging==23.0 +packaging==23.2 # via # -c .pin-constraints-torch.txt # huggingface-hub -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision @@ -103,7 +146,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -111,7 +154,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -r benchmarks/timm/requirements.in # huggingface-hub @@ -120,47 +163,46 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # huggingface-hub # torchvision -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir -safetensors==0.3.0 +safetensors==0.4.2 # via -r benchmarks/timm/requirements.in six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -torch==2.0.0+cu118 +torch==2.2.0+cu118 # via # -r benchmarks/timm/requirements.in # torchvision - # triton -torchvision==0.15.1+cu118 +torchvision==0.17.0+cu118 # via -r benchmarks/timm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin-constraints-torch.txt # huggingface-hub -triton==2.0.0 +triton==2.2.0 # via # -c .pin-constraints-torch.txt # torch -typing-extensions==4.5.0 +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # huggingface-hub # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/timm/requirements.rocm.txt b/benchmarks/timm/requirements.rocm.txt index 33980fa4d..2ce85b423 100644 --- a/benchmarks/timm/requirements.rocm.txt +++ b/benchmarks/timm/requirements.rocm.txt @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,36 +30,37 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 + # via + # -c .pin-constraints-torch.txt + # huggingface-hub + # torch + # triton +fsspec==2023.10.0 # via # -c .pin-constraints-torch.txt # huggingface-hub - # pytorch-triton-rocm # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -huggingface-hub==0.13.3 +huggingface-hub==0.20.3 # via -r benchmarks/timm/requirements.in -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -75,14 +72,66 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -c .pin-constraints-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -91,11 +140,11 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -packaging==23.0 +packaging==23.2 # via # -c .pin-constraints-torch.txt # huggingface-hub -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision @@ -103,7 +152,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -111,11 +160,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pytorch-triton-rocm==2.0.1 - # via - # -c .pin-constraints-torch.txt - # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -r benchmarks/timm/requirements.in # huggingface-hub @@ -124,43 +169,46 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # huggingface-hub # torchvision -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir -safetensors==0.3.0 +safetensors==0.4.2 # via -r benchmarks/timm/requirements.in six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -torch==2.0.0+rocm5.4.2 +torch==2.2.0 # via # -r benchmarks/timm/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.15.1+rocm5.4.2 +torchvision==0.17.0 # via -r benchmarks/timm/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via # -c .pin-constraints-torch.txt # huggingface-hub -typing-extensions==4.5.0 +triton==2.2.0 + # via + # -c .pin-constraints-torch.txt + # torch +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # huggingface-hub # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/torchvision/requirements.cuda.txt b/benchmarks/torchvision/requirements.cuda.txt index da06be073..2983e4b6e 100644 --- a/benchmarks/torchvision/requirements.cuda.txt +++ b/benchmarks/torchvision/requirements.cuda.txt @@ -2,7 +2,7 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=benchmarks/torchvision/requirements.cu118.txt --resolver=backtracking .pin-constraints-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision/requirements.cuda.txt --resolver=backtracking .pin-constraints-torchvision.txt benchmarks/torchvision/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # triton codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,33 +30,33 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 # via # -c .pin-constraints-torch.txt # torch # triton +fsspec==2023.10.0 + # via + # -c .pin-constraints-torch.txt + # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # triton -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -72,14 +68,60 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -c .pin-constraints-torch.txt # torchvision +nvidia-cublas-cu11==11.11.3.6 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu11 + # nvidia-cusolver-cu11 + # torch +nvidia-cuda-cupti-cu11==11.8.87 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu11==11.8.89 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu11==8.7.0.84 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu11==10.9.0.58 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu11==10.3.0.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu11==11.4.1.48 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu11==11.7.5.86 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nccl-cu11==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvtx-cu11==11.8.86 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -88,7 +130,7 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision @@ -96,7 +138,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -104,7 +146,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # omegaconf @@ -112,11 +154,11 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # torchvision -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir @@ -124,29 +166,28 @@ six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -torch==2.0.0+cu118 +torch==2.2.0+cu118 # via # -r benchmarks/torchvision/requirements.in # torchvision - # triton -torchvision==0.15.1+cu118 +torchvision==0.17.0+cu118 # via -r benchmarks/torchvision/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via -r benchmarks/torchvision/requirements.in -triton==2.0.0 +triton==2.2.0 # via # -c .pin-constraints-torch.txt # torch -typing-extensions==4.5.0 +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests diff --git a/benchmarks/torchvision/requirements.rocm.txt b/benchmarks/torchvision/requirements.rocm.txt index 639e07293..f8669923d 100644 --- a/benchmarks/torchvision/requirements.rocm.txt +++ b/benchmarks/torchvision/requirements.rocm.txt @@ -10,22 +10,18 @@ antlr4-python3-runtime==4.9.3 # via # -c .pin-constraints-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin-constraints-torch.txt # giving -certifi==2022.12.7 +certifi==2024.2.2 # via # -c .pin-constraints-torch.txt # requests -charset-normalizer==3.1.0 +charset-normalizer==3.3.2 # via # -c .pin-constraints-torch.txt # requests -cmake==3.26.1 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm codefind==0.1.3 # via # -c .pin-constraints-torch.txt @@ -34,33 +30,33 @@ executing==1.2.0 # via # -c .pin-constraints-torch.txt # varname -filelock==3.10.7 +filelock==3.13.1 + # via + # -c .pin-constraints-torch.txt + # torch + # triton +fsspec==2023.10.0 # via # -c .pin-constraints-torch.txt - # pytorch-triton-rocm # torch giving==0.4.2 # via # -c .pin-constraints-torch.txt # ptera # voir -idna==3.4 +idna==3.6 # via # -c .pin-constraints-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.3 # via # -c .pin-constraints-torch.txt # torch -lit==16.0.0 - # via - # -c .pin-constraints-torch.txt - # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin-constraints-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin-constraints-torch.txt # jinja2 @@ -72,14 +68,66 @@ mpmath==1.3.0 # via # -c .pin-constraints-torch.txt # sympy -networkx==3.1 +networkx==3.2.1 # via # -c .pin-constraints-torch.txt # torch -numpy==1.24.2 +numpy==1.26.3 # via # -c .pin-constraints-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin-constraints-torch.txt @@ -88,7 +136,7 @@ ovld==0.3.2 # via # -c .pin-constraints-torch.txt # voir -pillow==9.5.0 +pillow==10.2.0 # via # -c .pin-constraints-torch.txt # torchvision @@ -96,7 +144,7 @@ ptera==1.4.1 # via # -c .pin-constraints-torch.txt # voir -pygments==2.14.0 +pygments==2.17.2 # via # -c .pin-constraints-torch.txt # rich @@ -104,11 +152,7 @@ pynvml==11.5.0 # via # -c .pin-constraints-torch.txt # voir -pytorch-triton-rocm==2.0.1 - # via - # -c .pin-constraints-torch.txt - # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin-constraints-torch.txt # omegaconf @@ -116,11 +160,11 @@ reactivex==4.0.4 # via # -c .pin-constraints-torch.txt # giving -requests==2.28.2 +requests==2.31.0 # via # -c .pin-constraints-torch.txt # torchvision -rich==13.3.3 +rich==13.7.0 # via # -c .pin-constraints-torch.txt # voir @@ -128,25 +172,28 @@ six==1.16.0 # via # -c .pin-constraints-torch.txt # asttokens -sympy==1.11.1 +sympy==1.12 # via # -c .pin-constraints-torch.txt # torch -torch==2.0.0+rocm5.4.2 +torch==2.2.0 # via # -r benchmarks/torchvision/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.15.1+rocm5.4.2 +torchvision==0.17.0 # via -r benchmarks/torchvision/requirements.in -tqdm==4.65.0 +tqdm==4.66.1 # via -r benchmarks/torchvision/requirements.in -typing-extensions==4.5.0 +triton==2.2.0 + # via + # -c .pin-constraints-torch.txt + # torch +typing-extensions==4.9.0 # via # -c .pin-constraints-torch.txt # reactivex # torch -urllib3==1.26.15 +urllib3==1.26.18 # via # -c .pin-constraints-torch.txt # requests From e6814063a936df7819a3ed11712fd590923c5099 Mon Sep 17 00:00:00 2001 From: Setepenre Date: Mon, 26 Feb 2024 14:33:49 -0500 Subject: [PATCH 8/9] Tweak Performance computation (#144) (#203) Co-authored-by: Pierre Delaunay Co-authored-by: Olivier Breuleux --- milabench/_version.py | 5 + milabench/cli.py | 2 - milabench/compare.py | 15 +- milabench/report.py | 128 ++++++++++++------ milabench/summary.py | 8 ++ tests/config/benchio.yaml | 1 + .../benchio.0.R0.data | 3 +- .../benchio.0.R1.data | 2 +- .../benchio.1.R0.data | 2 +- .../benchio.1.R1.data | 2 +- .../benchio.0.data | 3 +- .../benchio.1.data | 3 +- tests/test_error.py | 1 - tests/test_summary.py | 16 ++- tests/test_summary/test_report.txt | 9 +- .../test_report_folder_does_average.txt | 9 +- 16 files changed, 149 insertions(+), 60 deletions(-) create mode 100644 milabench/_version.py diff --git a/milabench/_version.py b/milabench/_version.py new file mode 100644 index 000000000..0d2c1c57d --- /dev/null +++ b/milabench/_version.py @@ -0,0 +1,5 @@ +"""This file is generated, do not modify""" + +__tag__ = "v0.0.6-22-g3d7cb60" +__commit__ = "3d7cb6057452dfe0995fda3ab206eb4fef68952e" +__date__ = "2023-06-12 11:37:45 -0400" diff --git a/milabench/cli.py b/milabench/cli.py index d9e48b972..ccc43f680 100644 --- a/milabench/cli.py +++ b/milabench/cli.py @@ -337,7 +337,6 @@ def run(): if report: runs = {pack.logdir for pack in mp.packs.values()} - weights = None compare = None compare_gpus = False html = None @@ -351,7 +350,6 @@ def run(): make_report( summary, compare=compare, - weights=weights, html=html, compare_gpus=compare_gpus, price=price, diff --git a/milabench/compare.py b/milabench/compare.py index e2f23b2cb..e3b88b10c 100644 --- a/milabench/compare.py +++ b/milabench/compare.py @@ -11,6 +11,16 @@ class _Output: summary: dict = None +def retrieve_datetime_from_name(date): + """Windows does not support : in folders/filenames""" + formats = ["%Y-%m-%d_%H:%M:%S.%f", "%Y-%m-%d_%H_%M_%S.%f"] + for fmt in formats: + try: + return datetime.strptime(date, fmt) + except ValueError: + pass + + def fetch_runs(folder): runs = [] for run in os.listdir(folder): @@ -19,10 +29,13 @@ def fetch_runs(folder): continue if "." in run: name, date = run.split(".", maxsplit=1) - date = datetime.strptime(date, "%Y-%m-%d_%H:%M:%S.%f") + date = retrieve_datetime_from_name(date) else: name = run + + if date is None: date = datetime.fromtimestamp(os.path.getmtime(pth)) + out = _Output(pth, name, date) runs.append(out) diff --git a/milabench/report.py b/milabench/report.py index 5053763f6..7c369383c 100644 --- a/milabench/report.py +++ b/milabench/report.py @@ -17,17 +17,15 @@ def _make_row(summary, compare, weights): mkey = "train_rate" metric = "mean" row = {} - if weights is not None: - row["weight"] = weights["weight"] if weights else nan + row["n"] = summary["n"] if summary else nan row["fail"] = summary["failures"] if summary else nan row["perf"] = summary[mkey][metric] if summary else nan - row["perf_adj"] = ( - summary[mkey][metric] * (1 - row["fail"] / row["n"]) if summary else nan - ) + if compare: row["perf_base"] = compare[mkey][metric] row["perf_ratio"] = row["perf_adj"] / row["perf_base"] + row["std%"] = summary[mkey]["std"] / summary[mkey][metric] if summary else nan row["sem%"] = summary[mkey]["sem"] / summary[mkey][metric] if summary else nan # row["iqr%"] = (summary[mkey]["q3"] - summary[mkey]["q1"]) / summary[mkey]["median"] if summary else nan @@ -40,6 +38,23 @@ def _make_row(summary, compare, weights): if summary else nan ) + + # Sum of all the GPU performance + # to get the overall perf of the whole machine + if "per_gpu" in summary: + acc = 0 + for _, metrics in summary["per_gpu"].items(): + acc += metrics[metric] + else: + acc = row["perf"] + + success_ratio = 1 - row["fail"] / row["n"] + score = (acc if acc > 0 else row["perf"]) * success_ratio + + row["score"] = score + row["weight"] = weights.get("weight", summary.get("weight", 0)) + # ---- + return row @@ -157,21 +172,22 @@ def _report_pergpu(entries, measure="50"): return df -def make_report( - summary, - compare=None, - weights=None, - html=None, - compare_gpus=False, - price=None, - title=None, - sources=None, - errdata=None, -): - if weights: - weights = { - name: value for name, value in weights.items() if value.get("weight", 0) - } +columns_order = { + "fail": 1, + "n": 2, + "perf": 3, + "perf_adj": 4, + "sem%": 5, + "std%": 6, + "peak_memory": 7, + "score": 8, + "weight": 9, +} + + +def make_dataframe(summary, compare=None, weights=None): + if weights is None: + weights = dict() all_keys = list( sorted( @@ -193,6 +209,29 @@ def make_report( for key in all_keys } ).transpose() + + return df + + +@error_guard({}) +def make_report( + summary, + compare=None, + html=None, + compare_gpus=False, + price=None, + title=None, + sources=None, + errdata=None, + weights=None, +): + if weights is None: + weights = dict() + + df = make_dataframe(summary, compare, weights) + + # Reorder columns + df = df[sorted(df.columns, key=lambda k: columns_order.get(k, 0))] out = Outputter(stdout=sys.stdout, html=html) @@ -204,30 +243,31 @@ def make_report( out.title(title or "Benchmark results") out.print(df) - if weights: - out.section("Scores") - - def _score(column): - # This computes a weighted geometric mean - perf = df[column] - weights = df["weight"] - logscore = np.sum(np.log(perf) * weights) / np.sum(weights) - return np.exp(logscore) - - score = _score("perf") - failure_rate = df["fail"].sum() / df["n"].sum() - scores = { - "Failure rate": PassFail(failure_rate, failure_rate <= 0.01), - "Score": WithClass(f"{score:10.2f}", "score"), - } - if compare: - score_base = _score("perf_base") - scores.update({"Score (baseline)": score_base, "Ratio": score / score_base}) - if price: - rpp = price / score - scores["Price ($)"] = f"{price:10.2f}" - scores["RPP (Price / Score)"] = WithClass(f"{rpp:10.2f}", "rpp") - out.print(Table(scores)) + out.section("Scores") + + def _score(column): + # This computes a weighted geometric mean + perf = df[column] + weights = df["weight"] + logscore = np.sum(np.log(perf) * weights) / np.sum(weights) + return np.exp(logscore) + + score = _score("score") + failure_rate = df["fail"].sum() / df["n"].sum() + scores = { + "Failure rate": PassFail(failure_rate, failure_rate <= 0.01), + "Score": WithClass(f"{score:10.2f}", "score"), + } + if compare: + score_base = _score("perf_base") + scores.update({"Score (baseline)": score_base, "Ratio": score / score_base}) + + if price: + rpp = price / score + scores["Price ($)"] = f"{price:10.2f}" + scores["RPP (Price / Score)"] = WithClass(f"{rpp:10.2f}", "rpp") + + out.print(Table(scores)) if compare_gpus: for measure in ["mean", "min", "max"]: diff --git a/milabench/summary.py b/milabench/summary.py index e56dd1d45..f64fbc7d7 100644 --- a/milabench/summary.py +++ b/milabench/summary.py @@ -8,6 +8,7 @@ @error_guard(None) def aggregate(run_data): + """Group all the data inside a dictionary of lists""" omnibus = defaultdict(list) config = None start = None @@ -79,6 +80,7 @@ def aggregate(run_data): def _classify(all_aggregates): + """Group data by benchmark names""" classified = defaultdict(list) for agg in all_aggregates: config = agg["config"] @@ -87,12 +89,16 @@ def _classify(all_aggregates): def _merge(aggs): + """Merge same bench data into a single list of observations""" + results = {"data": defaultdict(list)} for agg in aggs: data = agg.pop("data") results.update(agg) + for k, v in data.items(): results["data"][k].extend(v) + return results @@ -128,6 +134,7 @@ def _metrics(xs): def _summarize(group): agg = group["data"] gpudata = defaultdict(lambda: defaultdict(list)) + for entry in agg["gpudata"]: for device, data in entry.items(): if data["memory"][0] == 1 or data["load"] == 0: @@ -157,6 +164,7 @@ def _summarize(group): } for device, data in gpudata.items() }, + "weight": config.get("weight", 0), } diff --git a/tests/config/benchio.yaml b/tests/config/benchio.yaml index 960d4a756..9f74afa99 100644 --- a/tests/config/benchio.yaml +++ b/tests/config/benchio.yaml @@ -1,6 +1,7 @@ benchio: inherits: _defaults definition: ../yoshua-benchio + weight: 2 plan: method: njobs n: 2 diff --git a/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.0.R0.data b/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.0.R0.data index 548d74b6e..022d63f2a 100644 --- a/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.0.R0.data +++ b/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.0.R0.data @@ -1,4 +1,5 @@ -{"event": "config", "data": {"dirs": {"code": "/Tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/Tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/Tmp/slurm.2998973.0/benchio/data", "runs": "/Tmp/slurm.2998973.0/benchio/runs", "extra": "/Tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/Tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "0", "R0"], "run_name": "rijubigo.2023-03-24_13:45:27.512446", "arch": "cuda", "install_variant": "", "job-number": 0}, "pipe": null} +{"event": "config", "data": {"weight": 1, "dirs": {"code": "/tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/tmp/slurm.2998973.0/benchio/data", "runs": "/tmp/slurm.2998973.0/benchio/runs", "extra": "/tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "0", "R0"], "run_name": "rijubigo.2023-03-24_13:45:27.512446", "arch": "cuda", "install_variant": "", "job-number": 0}, "pipe": null} +{"event": "meta", "data": {}, "pipe": "data"} {"event": "start", "data": {"command": ["voir", "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio/main.py"], "time": 1679679927.610199}, "pipe": null} {"event": "phase", "data": {"name": "init"}, "pipe": "data"} {"event": "phase", "data": {"name": "parse_args"}, "pipe": "data"} diff --git a/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.0.R1.data b/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.0.R1.data index 10969415f..4006a0e3d 100644 --- a/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.0.R1.data +++ b/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.0.R1.data @@ -1,4 +1,4 @@ -{"event": "config", "data": {"dirs": {"code": "/Tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/Tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/Tmp/slurm.2998973.0/benchio/data", "runs": "/Tmp/slurm.2998973.0/benchio/runs", "extra": "/Tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/Tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "0", "R1"], "run_name": "rijubigo.2023-03-24_13:45:27.512446", "arch": "cuda", "install_variant": "", "job-number": 0}, "pipe": null} +{"event": "config", "data": {"weight": 1, "dirs": {"code": "/tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/tmp/slurm.2998973.0/benchio/data", "runs": "/tmp/slurm.2998973.0/benchio/runs", "extra": "/tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "0", "R1"], "run_name": "rijubigo.2023-03-24_13:45:27.512446", "arch": "cuda", "install_variant": "", "job-number": 0}, "pipe": null} {"event": "start", "data": {"command": ["voir", "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio/main.py"], "time": 1679679928.498935}, "pipe": null} {"event": "phase", "data": {"name": "init"}, "pipe": "data"} {"event": "phase", "data": {"name": "parse_args"}, "pipe": "data"} diff --git a/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.1.R0.data b/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.1.R0.data index 49035aac3..36c6ccd2e 100644 --- a/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.1.R0.data +++ b/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.1.R0.data @@ -1,4 +1,4 @@ -{"event": "config", "data": {"dirs": {"code": "/Tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/Tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/Tmp/slurm.2998973.0/benchio/data", "runs": "/Tmp/slurm.2998973.0/benchio/runs", "extra": "/Tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/Tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "1", "R0"], "run_name": "rijubigo.2023-03-24_13:45:27.512446", "arch": "cuda", "install_variant": "", "job-number": 1}, "pipe": null} +{"event": "config", "data": {"weight": 1, "dirs": {"code": "/tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/tmp/slurm.2998973.0/benchio/data", "runs": "/tmp/slurm.2998973.0/benchio/runs", "extra": "/tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "1", "R0"], "run_name": "rijubigo.2023-03-24_13:45:27.512446", "arch": "cuda", "install_variant": "", "job-number": 1}, "pipe": null} {"event": "start", "data": {"command": ["voir", "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio/main.py"], "time": 1679679927.640314}, "pipe": null} {"event": "phase", "data": {"name": "init"}, "pipe": "data"} {"event": "phase", "data": {"name": "parse_args"}, "pipe": "data"} diff --git a/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.1.R1.data b/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.1.R1.data index d26e308a3..7f2674a56 100644 --- a/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.1.R1.data +++ b/tests/runs/rijubigo.2023-03-24_13:45:27.512446/benchio.1.R1.data @@ -1,4 +1,4 @@ -{"event": "config", "data": {"dirs": {"code": "/Tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/Tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/Tmp/slurm.2998973.0/benchio/data", "runs": "/Tmp/slurm.2998973.0/benchio/runs", "extra": "/Tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/Tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "1", "R1"], "run_name": "rijubigo.2023-03-24_13:45:27.512446", "arch": "cuda", "install_variant": "", "job-number": 1}, "pipe": null} +{"event": "config", "data": {"weight": 1, "dirs": {"code": "/tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/tmp/slurm.2998973.0/benchio/data", "runs": "/tmp/slurm.2998973.0/benchio/runs", "extra": "/tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "1", "R1"], "run_name": "rijubigo.2023-03-24_13:45:27.512446", "arch": "cuda", "install_variant": "", "job-number": 1}, "pipe": null} {"event": "start", "data": {"command": ["voir", "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio/main.py"], "time": 1679679928.529273}, "pipe": null} {"event": "phase", "data": {"name": "init"}, "pipe": "data"} {"event": "phase", "data": {"name": "parse_args"}, "pipe": "data"} diff --git a/tests/runs/sedumoje.2023-03-24_13:57:35.089747/benchio.0.data b/tests/runs/sedumoje.2023-03-24_13:57:35.089747/benchio.0.data index 0964160c2..50a85f734 100644 --- a/tests/runs/sedumoje.2023-03-24_13:57:35.089747/benchio.0.data +++ b/tests/runs/sedumoje.2023-03-24_13:57:35.089747/benchio.0.data @@ -1,4 +1,5 @@ -{"event": "config", "data": {"dirs": {"code": "/Tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/Tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/Tmp/slurm.2998973.0/benchio/data", "runs": "/Tmp/slurm.2998973.0/benchio/runs", "extra": "/Tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/Tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "0"], "run_name": "sedumoje.2023-03-24_13:57:35.089747", "arch": "cuda", "install_variant": "", "job-number": 0}, "pipe": null} +{"event": "config", "data": {"weight": 1, "dirs": {"code": "/tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/tmp/slurm.2998973.0/benchio/data", "runs": "/tmp/slurm.2998973.0/benchio/runs", "extra": "/tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "0"], "run_name": "sedumoje.2023-03-24_13:57:35.089747", "arch": "cuda", "install_variant": "", "job-number": 0}, "pipe": null} +{"event": "meta", "data": {}, "pipe": "data"} {"event": "start", "data": {"command": ["voir", "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio/main.py"], "time": 1679680655.1949446}, "pipe": null} {"event": "phase", "data": {"name": "init"}, "pipe": "data"} {"event": "phase", "data": {"name": "parse_args"}, "pipe": "data"} diff --git a/tests/runs/sedumoje.2023-03-24_13:57:35.089747/benchio.1.data b/tests/runs/sedumoje.2023-03-24_13:57:35.089747/benchio.1.data index 76fb0c0ce..9ca0d1d2c 100644 --- a/tests/runs/sedumoje.2023-03-24_13:57:35.089747/benchio.1.data +++ b/tests/runs/sedumoje.2023-03-24_13:57:35.089747/benchio.1.data @@ -1,4 +1,5 @@ -{"event": "config", "data": {"dirs": {"code": "/Tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/Tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/Tmp/slurm.2998973.0/benchio/data", "runs": "/Tmp/slurm.2998973.0/benchio/runs", "extra": "/Tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/Tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "1"], "run_name": "sedumoje.2023-03-24_13:57:35.089747", "arch": "cuda", "install_variant": "", "job-number": 1}, "pipe": null} +{"event": "config", "data": {"weight": 1, "dirs": {"code": "/tmp/slurm.2998973.0/benchio/code/benchio", "venv": "/tmp/slurm.2998973.0/benchio/venv/benchio", "data": "/tmp/slurm.2998973.0/benchio/data", "runs": "/tmp/slurm.2998973.0/benchio/runs", "extra": "/tmp/slurm.2998973.0/benchio/extra/benchio", "base": "/tmp/slurm.2998973.0/benchio"}, "venv": {"type": "virtualenv"}, "config_base": "/home/mila/b/breuleuo/code/nv-milabench/tests/config", "config_file": "/home/mila/b/breuleuo/code/nv-milabench/tests/config/benchio.yaml", "plan": {"method": "njobs", "n": 2}, "definition": "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio", "name": "benchio", "group": "benchio", "tag": ["benchio", "1"], "run_name": "sedumoje.2023-03-24_13:57:35.089747", "arch": "cuda", "install_variant": "", "job-number": 1}, "pipe": null} +{"event": "meta", "data": {}, "pipe": "data"} {"event": "start", "data": {"command": ["voir", "/home/mila/b/breuleuo/code/nv-milabench/tests/yoshua-benchio/main.py"], "time": 1679680655.2239232}, "pipe": null} {"event": "phase", "data": {"name": "init"}, "pipe": "data"} {"event": "phase", "data": {"name": "parse_args"}, "pipe": "data"} diff --git a/tests/test_error.py b/tests/test_error.py index 9b16b05f8..dbaeaf560 100644 --- a/tests/test_error.py +++ b/tests/test_error.py @@ -26,4 +26,3 @@ def test_error_reporting_short(capsys, args, matcher, config): captured = capsys.readouterr() assert matcher(captured.out), "The traceback need to be printed" - assert captured.err == "" diff --git a/tests/test_summary.py b/tests/test_summary.py index d9301e406..97d666691 100644 --- a/tests/test_summary.py +++ b/tests/test_summary.py @@ -1,10 +1,10 @@ from milabench.cli import main -def test_report(runs_folder, capsys, file_regression): +def test_report(runs_folder, capsys, file_regression, config): folder = runs_folder / "rijubigo.2023-03-24_13:45:27.512446" try: - main(["report", "--runs", folder]) + main(["report", "--runs", folder, "--config", config("benchio")]) except SystemExit as exc: assert not exc.code @@ -33,3 +33,15 @@ def test_compare(runs_folder, capsys, file_regression): output = capsys.readouterr().out output = output.replace(str(runs_folder), "XXX") file_regression.check(output) + + +def test_summary_full(runs_folder): + from milabench.cli import _read_reports, make_summary, make_report + + run = runs_folder / "rijubigo.2023-03-24_13:45:27.512446" + + runs = [run] + reports = _read_reports(*runs) + summary = make_summary(reports.values()) + + make_report(summary, None) diff --git a/tests/test_summary/test_report.txt b/tests/test_summary/test_report.txt index 9a214f7e1..7d8474e01 100644 --- a/tests/test_summary/test_report.txt +++ b/tests/test_summary/test_report.txt @@ -2,5 +2,10 @@ Source: XXX ================= Benchmark results ================= - fail n peak_memory perf perf_adj sem% std% -benchio 0 4 -1 7979.82 7979.82 2.9% 17.2% + fail n perf sem% std% peak_memory score weight +benchio 0 4 7979.82 2.9% 17.2% -1 7979.81831 2.00 + +Scores +------ +Failure rate: 0.00% (PASS) +Score: 7979.82 diff --git a/tests/test_summary/test_report_folder_does_average.txt b/tests/test_summary/test_report_folder_does_average.txt index 54feb56d6..1b9b00f4c 100644 --- a/tests/test_summary/test_report_folder_does_average.txt +++ b/tests/test_summary/test_report_folder_does_average.txt @@ -2,5 +2,10 @@ Source: XXX ================= Benchmark results ================= - fail n peak_memory perf perf_adj sem% std% -benchio 0 6 -1 7878.45 7878.45 2.5% 18.0% + fail n perf sem% std% peak_memory score weight +benchio 0 6 7878.45 2.5% 18.0% 24456 7878.451302 1.00 + +Scores +------ +Failure rate: 0.00% (PASS) +Score: 7878.45 From 77fa0bc82b305aac872d4804dbf33e2fe9a75e80 Mon Sep 17 00:00:00 2001 From: Setepenre Date: Tue, 27 Feb 2024 11:47:43 -0500 Subject: [PATCH 9/9] Use latest dataset (#204) --- benchmarks/accelerate_opt/main.py | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/benchmarks/accelerate_opt/main.py b/benchmarks/accelerate_opt/main.py index 61bf1fa1b..ec225c707 100644 --- a/benchmarks/accelerate_opt/main.py +++ b/benchmarks/accelerate_opt/main.py @@ -145,19 +145,20 @@ def mblog(data): validation_split_percentage = config["validation_split_percentage"] dataset_name = config["dataset_name"] dataset_config_name = config["dataset_config_name"] - raw_datasets = load_dataset(dataset_name, dataset_config_name, revision="f5562967961a45407fa15044c5535a607200983f") + revision = "b08601e04326c79dfdd32d625aee71d232d685c3" + raw_datasets = load_dataset(dataset_name, dataset_config_name, revision=revision) if "validation" not in raw_datasets.keys(): raw_datasets["validation"] = load_dataset( dataset_name, dataset_config_name, split=f"train[:{validation_split_percentage}%]", - revision="f5562967961a45407fa15044c5535a607200983f", + revision=revision, ) raw_datasets["train"] = load_dataset( dataset_name, dataset_config_name, split=f"train[{validation_split_percentage}%:]", - revision="f5562967961a45407fa15044c5535a607200983f", + revision=revision, ) model_name = config["model_name"]