-
Notifications
You must be signed in to change notification settings - Fork 52
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Fix] "auto_tuner" should be under the field config.experiment. #301
Conversation
…be under the field "runner" change no_shared_fs = config.experiment.get("no_shared_fs", False) to no_shared_fs = runner_config.get("no_shared_fs", False)
I want to change the default metric to TFLOPs and change the order to descend, but it doesn't work. Because, the "auto_tuner" is under config.experiment instead of config. After making the following changes it worked. Change if ( "auto_tuner" in self.config and "performance" in self.config.experiment.auto_tuner ): to if ( "auto_tuner" in self.config.experiment and "performance" in self.config.experiment.auto_tuner ):
Thank you for your valuable input! Your contribution is greatly appreciated. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
fix test_parallel_context.py fix ut [Fix] "auto_tuner" should be under the field config.experiment. (FlagOpen#301) I want to change the default metric to TFLOPs and change the order to descend, but it doesn't work. Because, the "auto_tuner" is under config.experiment instead of config. After making the following changes it worked. Change if ( "auto_tuner" in self.config and "performance" in self.config.experiment.auto_tuner ): to if ( "auto_tuner" in self.config.experiment and "performance" in self.config.experiment.auto_tuner ): add 'attention_backend: unfused' for functional tests update extra validation feature (FlagOpen#303) 1. usage: ```yaml extra_eval_interval: 5 extra_valid_data_path: [ weight1, data_path1, weight2, data_path2, ] ``` - `weight` refers to the number of tokens for extra validation in `data_path`. - **NOTE: The extra validation always starts from consumed_sample=0.** 2. output format ``` (min, max) time across ranks (ms): evaluate .......................................: (xxx, xxx) ------------------------------------------------------------------------------- extra validation iteration 5 loss at data_path1 | consumed samples: xxx | lm loss value: xxx | lm loss PPL: xxx | ------------------------------------------------------------------------------- (min, max) time across ranks (ms): evaluate .......................................: (xxx, xxx) ------------------------------------------------------------------------------- extra validation iteration 5 loss at data_path2 | consumed samples: xxx | lm loss value: xxx | lm loss PPL: xxx | ------------------------------------------------------------------------------- ``` Fix extra validation corner case (FlagOpen#304) polish train.py
@changsci Thank you for the two provided fix PRs. Are you still a student? If so, FlagScale also offers some internship opportunities. If you are interested, you can contact us via email: [email protected] |
Thanks for your letter. Currently, I am working for ZTE. It must be mentioned that this project has inspired me greatly. It is truly a remarkable project. I sincerely hope it will continue to thrive and be seen by more developers.
Thank you again for sharing this wonderful project!
…---Original---
From: ***@***.***>
Date: Fri, Jan 3, 2025 17:40 PM
To: ***@***.***>;
Cc: ***@***.******@***.***>;
Subject: Re: [FlagOpen/FlagScale] [Fix] "auto_tuner" should be under the fieldconfig.experiment. (PR #301)
@changsci Thank you for the two provided fix PRs. Are you still a student? If so, FlagScale also offers some internship opportunities. If you are interested, you can contact us via email: ***@***.***
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you were mentioned.Message ID: ***@***.***>
|
@changsci Thank you so much. I really appreciate it. Looking forward to more exchanges and let's keep in touch. |
I want to change the default metric to TFLOPs and change the order to descend, but it doesn't work. Because, the "auto_tuner" is under config.experiment instead of config.
After making the following changes it worked.
Change
if (
"auto_tuner" in self.config
and "performance" in self.config.experiment.auto_tuner
):
to
if (
"auto_tuner" in self.config.experiment
and "performance" in self.config.experiment.auto_tuner
):