Parsing Codalab #112
-
First, I want to thank you for creating WILDS, it is a great addition to the landscape of datasets for ML. I would like to make use of the models and data in CodaLab, but I am a little lost. Is there a document that describes the directories and files in Codalab? What are all the different files? How can I reproduce the results? For example, I don't yet quite understand what is ERM in this context? Does it mean SGD on an approximation to the average error? Thanks, |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Hi Yoav, Thank you for your kind words and your interest in WILDS!
Here are the relevant file outputs from each experiment:
We provide a set of scripts that can be used to reproduce the benchmark results reported in the WILDS papers. You can find the scripts at https://github.com/p-lambda/wilds, and the README has additional instructions (e.g., https://github.com/p-lambda/wilds#using-the-example-scripts). So once you find the exact commands and hyperparameters on the Codalab worksheets, you can run these commands using these scripts.
By ERM, we refer to the standard training procedure where we minimize the average loss. The optimizer used depends on the dataset. Hope this helps, and let us know if you have any additional questions! Best, |
Beta Was this translation helpful? Give feedback.
Hi Yoav,
Thank you for your kind words and your interest in WILDS!
Here are the relevant file outputs from each experiment:
{split}_eval.csv
: Reports model performance on the specified data split at the end of each epoch. The reported metrics include the official evaluation metrics for each dataset (e.g., macro F1 for the iWildCam dataset).{split}_algo.csv
: Reports additional metrics used to monitor training (e.g., loss).best_model.pth
: Saved weights for the model with best validation performance, as measured by the official evaluation metrics for each dataset. More specifically…