Skip to content

Commit

Permalink
Use universal downloader for bert-base (#780)
Browse files Browse the repository at this point in the history
  • Loading branch information
gfursin authored Jun 1, 2023
2 parents 6805473 + 61bad16 commit 66b4306
Show file tree
Hide file tree
Showing 36 changed files with 155 additions and 118 deletions.
37 changes: 37 additions & 0 deletions cm-mlops/automation/list_of_scripts.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,37 @@
[ [Back to index](README.md) ]

<!--
This file is generated automatically - don't edit!
-->

This is an automatically generated list of reusable CM scripts being developed
by the [open taskforce on automation and reproducibility](https://github.com/mlcommons/ck/issues/536)
to make MLOps and DevOps tools more interoperable, portable, deterministic and reproducible.
These scripts suppport the community effort to modularize ML Systems and automate their bechmarking, optimization,
design space exploration and deployment across continuously changing software and hardware.

# List of CM scripts by categories

<details>
<summary>Click here to see the table of contents.</summary>

* [Platform information](#platform-information)


</details>

### Platform information

* [detect-os](https://github.com/mlcommons/ck/tree/master/cm-mlops/script/detect-os)


# List of all sorted CM scripts

* [detect-os](https://github.com/mlcommons/ck/tree/master/cm-mlops/script/detect-os)




# Maintainers

* [Open MLCommons taskforce on automation and reproducibility](https://github.com/mlcommons/ck/blob/master/docs/taskforce.md)'
4 changes: 2 additions & 2 deletions cm-mlops/automation/script/README-extra.md
Original file line number Diff line number Diff line change
Expand Up @@ -708,12 +708,12 @@ as shown in the next example.
Instead of adding this flag to all scripts, you can specify it
using `CM_SCRIPT_EXTRA_CMD` environment variable as follows:
```bash
export CM_SCRIPT_EXTRA_CMD="--adr.python.name.mlperf"
export CM_SCRIPT_EXTRA_CMD="--adr.python.name=mlperf"
```

You can even specify min Python version required as follows:
```bash
export CM_SCRIPT_EXTRA_CMD="--adr.python.name.mlperf --adr.python.version_min=3.9"
export CM_SCRIPT_EXTRA_CMD="--adr.python.name=mlperf --adr.python.version_min=3.9"
```

### Assembling pipelines with other artifacts included
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ Under preparation: Reproduce and optimize MLPerf inference benchmarks during Stu

See our [related challange from 2022](https://access.cknowledge.org/playground/?action=challenges&name=optimize-mlperf-inference-scc2023).

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cKnowledge](https://cKnowledge.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
Prepare, optimize and reproduce MLPerf inference v2.1 benchmarks across diverse implementations, software and hardware
using the [MLCommons CK framework](https://github.com/mlcommons/ck).

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ using the [MLCommons CK framework](https://github.com/mlcommons/ck):
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
10 changes: 7 additions & 3 deletions cm-mlops/challenge/optimize-mlperf-inference-v3.1-2023/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,9 +3,10 @@
Prepare, optimize and reproduce MLPerf inference v3.1 benchmarks across diverse implementations, models, software and hardware.

Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to run and optimize MLPerf inference benchmarks.
how to use and enhance [CM scripts and workflows](https://github.com/mlcommons/ck/blob/master/docs/README.md)
to run and optimize MLPerf inference benchmarks on your software/hardware stack.

## Organizers
#### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand All @@ -22,6 +23,9 @@ For MLPerf inference 3.1 we have the following benchmark tasks
6. Recommendation using DLRM model and Criteo dataset
7. Large Language Model (Pending)

All the six tasks are applicable to datacenter while all except Recommendation are applicable to edge category. Further, language processing and medical imaging models have a high accuracy variant where the achieved accuracy must be within `99.9%` (`99%` is the default accuracy requirement) of the fp32 reference model. Recommendation task is only having a high accuracy variant. Currently we are not supporting Recommendation task as we are not having a highend server which is a requirement.
All the six tasks are applicable to datacenter while all except Recommendation are applicable to edge category.
Further, language processing and medical imaging models have a high accuracy variant where the achieved accuracy
must be within `99.9%` (`99%` is the default accuracy requirement) of the fp32 reference model.
Recommendation task is only having a high accuracy variant. Currently we are not supporting Recommendation task as we are not having a highend server which is a requirement.

This challenge is integrated with [our platform](https://github.com/ctuning/mlcommons-ck/tree/master/platform)
Original file line number Diff line number Diff line change
@@ -1,11 +1,19 @@
## Setup
Please follow the MLCommons CK [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md) to install CM.
Download the ck repo to get the CM script for MLPerf submission

Please follow this [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md)
to install the MLCommons CM reproducibility and automation language in your native environment or Docker container.

Then install the repository with CM automation scripts to run MLPerf benchmarks out-of-the-box
across different software, hardware, models and data sets:


```
cm pull repo mlcommons@ck
```

Note that you can install Python virtual environment via CM to avoid contaminating
your local Python installation as described [here](https://github.com/mlcommons/ck/blob/master/cm-mlops/automation/script/README-extra.md#using-python-virtual-environments).

## Run Commands

3d-unet has two variants - `3d-unet-99` and `3d-unet-99.9` where the `99` and `99.9` specifies the required accuracy constraint with respect to the reference floating point model. Both models can be submitter under edge as well as datacenter category.
Expand Down
Original file line number Diff line number Diff line change
@@ -1,15 +1,19 @@
## Setup

Please follow this [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md)
to install the [MLCommons CM scripting language](https://github.com/mlcommons/ck/tree/master/docs#collective-mind-language-cm)
on your system with minimal dependencies.
to install the MLCommons CM reproducibility and automation language in your native environment or Docker container.

Then install the repository with CM automation scripts to run MLPerf benchmarks out-of-the-box
across different software, hardware, models and data sets:

Download a GitHub repository with [portable and reusable CM scripts](https://github.com/mlcommons/ck/tree/master/cm-mlops/script)
for unified MLPerf benchmarking and submission:

```
cm pull repo mlcommons@ck
```

Note that you can install Python virtual environment via CM to avoid contaminating
your local Python installation as described [here](https://github.com/mlcommons/ck/blob/master/cm-mlops/automation/script/README-extra.md#using-python-virtual-environments).

## Run Commands

Bert has two variants - `bert-99` and `bert-99.9` where the `99` and `99.9` specifies the required accuracy constraint with respect to the reference floating point model. `bert-99.9` model is applicable only on a datacenter system.
Expand Down
Original file line number Diff line number Diff line change
@@ -1,11 +1,19 @@
## Setup
Please follow the MLCommons CK [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md) to install CM.
Download the ck repo to get the CM script for MLPerf submission

Please follow this [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md)
to install the MLCommons CM reproducibility and automation language in your native environment or Docker container.

Then install the repository with CM automation scripts to run MLPerf benchmarks out-of-the-box
across different software, hardware, models and data sets:


```
cm pull repo mlcommons@ck
```

Note that you can install Python virtual environment via CM to avoid contaminating
your local Python installation as described [here](https://github.com/mlcommons/ck/blob/master/cm-mlops/automation/script/README-extra.md#using-python-virtual-environments).

## Run Commands

We need to get imagenet full dataset to make image-classification submissions for MLPerf inference. Since this dataset is not publicly available via a URL please follow the instructions given [here](https://github.com/mlcommons/ck/blob/master/cm-mlops/script/get-dataset-imagenet-val/README-extra.md) to download the dataset and register in CM.
Expand Down
Original file line number Diff line number Diff line change
@@ -1,11 +1,19 @@
## Setup
Please follow the MLCommons CK [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md) to install CM.
Download the ck repo to get the CM script for MLPerf submission

Please follow this [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md)
to install the MLCommons CM reproducibility and automation language in your native environment or Docker container.

Then install the repository with CM automation scripts to run MLPerf benchmarks out-of-the-box
across different software, hardware, models and data sets:


```
cm pull repo mlcommons@ck
```

Note that you can install Python virtual environment via CM to avoid contaminating
your local Python installation as described [here](https://github.com/mlcommons/ck/blob/master/cm-mlops/automation/script/README-extra.md#using-python-virtual-environments).

## Run Commands


Expand Down
Original file line number Diff line number Diff line change
@@ -1,11 +1,19 @@
## Setup
Please follow the MLCommons CK [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md) to install CM.
Download the ck repo to get the CM script for MLPerf submission

Please follow this [installation guide](https://github.com/mlcommons/ck/blob/master/docs/installation.md)
to install the MLCommons CM reproducibility and automation language in your native environment or Docker container.

Then install the repository with CM automation scripts to run MLPerf benchmarks out-of-the-box
across different software, hardware, models and data sets:


```
cm pull repo mlcommons@ck
```

Note that you can install Python virtual environment via CM to avoid contaminating
your local Python installation as described [here](https://github.com/mlcommons/ck/blob/master/cm-mlops/automation/script/README-extra.md#using-python-virtual-environments).

## Run Commands

### TensorRT backend
Expand Down
Original file line number Diff line number Diff line change
@@ -1,3 +1,5 @@
## Setup ASW instance for MLPerf

The below instructions are for creating an AWS instance from the CLI. You can also create an instance via web and setup CM on it.

## Prerequisites
Expand Down
Original file line number Diff line number Diff line change
@@ -1,3 +1,5 @@
## Setup GCP instance for MLPerf

The below instructions are for creating a Google Cloud instance from the CLI. You can also create an instance via web and setup CM on it.

## Prerequisites
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
## Setup

We used Nvidia Jetson AGX Orin developer kit with 32GB RAM and 64GB eMMC. We also connected a 500GB SSD disk via USB and Wifi connection for internet connectivity.

We used the out of the box developer kit image which was running Ubuntu 20.04 and JetPack 5.0.1 Developer Preview (L4T 34.1.1) with CUDA 11.4. We were also using the default 4k page size (Nvidia recommends 64k for MLPerf inference).
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Prepare and optimize MLPerf inference v3.1 submission for publicly-available Ama
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Prepare and optimize MLPerf inference v3.1 submission for AMD-based platforms.
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Prepare and optimize MLPerf inference v3.1 submission for AMD-based platforms.
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Prepare and optimize MLPerf inference v3.1 submission for publicly-available Goo
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ MLPerf BERT model is available at Hugging Face [here](https://huggingface.co/ctu
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Prepare and optimize MLPerf inference v3.1 submission for Intel-based platforms.
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ using KILT. Compare usability with MLCommons MITL.
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Compare usability and results with KILT.
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Prepare and optimize MLPerf inference v3.1 submission for Nvidia GPUs.
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Prepare and optimize MLPerf inference v3.1 submission for Qualcomm AI100-based p
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ using the [MLCommons CK framework](https://github.com/mlcommons/ck):
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to run and optimize MLPerf inference benchmarks.

## Organizers
### Organizers

* [Deelvin](https://deelvin.com)
* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
Make it possible to run MLPerf inference v3.1 benchmarks on Windows
using the [MLCommons CK framework](https://github.com/mlcommons/ck).

## Organizers
### Organizers

* Stanley Mwangi (Microsoft)
* Grigori Fursin (MLCommons, cTuning & cKnowledge)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ Join this public [Discord server](https://discord.gg/JjWNWXKxwT)
to discuss with the community and organizers
how to use and enhance CK to run and optimize MLPerf inference benchmarks.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ Join this public [Discord server](https://discord.gg/JjWNWXKxwT)
to discuss with the community and organizers
how to use and enhance CK to run and optimize MLPerf inference benchmarks.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
2 changes: 1 addition & 1 deletion cm-mlops/challenge/repro-mlperf-inf-v3.0-orin/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ Reproduce MLPerf inference v3.0 benchmark results for Nvidia Jetson Orin
Join this public [Discord server](https://discord.gg/JjWNWXKxwT) to discuss with the community and organizers
how to use and enhance CK to benchmark and optimize ML Systems.

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
Reproduce the MLPerf inference RetinaNet benchmark during Student Cluster Competition at SuperComputing'22
using the following [CK2(CM) tutorial](https://github.com/mlcommons/ck/blob/master/docs/tutorials/sc22-scc-mlperf.md).

## Organizers
### Organizers

* [MLCommons taskforce on automation and reproducibility](https://cKnowledge.org/mlcommons-taskforce)
* [cTuning foundation](https://cTuning.org)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@

Reproduce and automate IPOL paper (proof-of-concept):

### Organizers
#### Organizers

* Jose Hernandez
* Miguel Colom
Expand Down
Loading

0 comments on commit 66b4306

Please sign in to comment.