diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index 395bbd126..b46ede5ff 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -82,7 +82,6 @@ filelock==3.15.4 # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub - # pytorch-triton-rocm # torch # transformers frozenlist==1.4.1 diff --git a/benchmarks/huggingface/requirements.rocm.txt b/benchmarks/huggingface/requirements.rocm.txt index 1f54d841a..54dd45ae9 100644 --- a/benchmarks/huggingface/requirements.rocm.txt +++ b/benchmarks/huggingface/requirements.rocm.txt @@ -34,7 +34,6 @@ filelock==3.15.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub - # pytorch-triton-rocm # torch # transformers fsspec==2024.6.1 @@ -84,6 +83,58 @@ numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt diff --git a/benchmarks/timm/requirements.rocm.txt b/benchmarks/timm/requirements.rocm.txt index 8383f9e6b..61ec4d6a6 100644 --- a/benchmarks/timm/requirements.rocm.txt +++ b/benchmarks/timm/requirements.rocm.txt @@ -34,7 +34,6 @@ filelock==3.15.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub - # pytorch-triton-rocm # torch fsspec==2024.6.1 # via @@ -82,6 +81,58 @@ numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt diff --git a/benchmarks/torchvision/requirements.rocm.txt b/benchmarks/torchvision/requirements.rocm.txt index 094eb29b6..f7004b571 100644 --- a/benchmarks/torchvision/requirements.rocm.txt +++ b/benchmarks/torchvision/requirements.rocm.txt @@ -68,6 +68,58 @@ numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin-constraints-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.19.3 + # via + # -c .pin-constraints-torch.txt + # torch +nvidia-nvjitlink-cu12==12.3.101 + # via + # -c .pin-constraints-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin-constraints-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt diff --git a/milabench/report.py b/milabench/report.py index aebcaf093..e87ac380b 100644 --- a/milabench/report.py +++ b/milabench/report.py @@ -272,6 +272,29 @@ def sort_by(key): for key in all_keys } ).transpose() + + return df + + +@error_guard({}) +def make_report( + summary, + compare=None, + html=None, + compare_gpus=False, + price=None, + title=None, + sources=None, + errdata=None, + weights=None, +): + if weights is None: + weights = dict() + + df = make_dataframe(summary, compare, weights) + + # Reorder columns + df = df[sorted(df.columns, key=lambda k: columns_order.get(k, 0))] # Reorder columns df = df[sorted(df.columns, key=lambda k: columns_order.get(k, 2000))]