-
Notifications
You must be signed in to change notification settings - Fork 287
Add optimization level for tuning strategy #179
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
84ae658 to
52825cd
Compare
| # self.best_tune_result = self.baseline | ||
| # Initialize the best qmodel as fp32 model | ||
| self.best_qmodel = self._fp32_model | ||
| # self.best_qmodel = self._fp32_model |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
why this line gets commented off?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We will initialize the self.best_qmodel with the first tuning result whose accuracy meets the requirements. code
If we initialize the self.best_qmodel with the fp32 model, we may return the fp32 model as the tuning result which may confuse the user, right?
Signed-off-by: chensuyue <[email protected]> Co-authored-by: WenjiaoYue <[email protected]> Signed-off-by: yiliu30 <[email protected]>
* Fix Time_series_LSTM tuning issue Signed-off-by: Lv, Liang1 <[email protected]> * fix pylint issue Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Li, Chenda <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: chensuyue <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Wang, Chang1 <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Sun, Xuehao <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Li, Chenda <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: intel-zhangyi <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Wang, Chang1 <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: chensuyue <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Xinyu Ye <[email protected]> Signed-off-by: yiliu30 <[email protected]>
…166) Signed-off-by: Xinyu Ye <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: changwa1 <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: yiliu30 <[email protected]>
* Fix centernet_hg104 tuning issue Signed-off-by: sys-lpot-val <[email protected]> * Fix TextRNN tuning issue Signed-off-by: sys-lpot-val <[email protected]> Signed-off-by: sys-lpot-val <[email protected]> Signed-off-by: Lv, Liang1 <[email protected]> Co-authored-by: sys-lpot-val <[email protected]> Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
|
The branch was polluted, the feature will be added by PR 227, close it first. |
* refine PatchVLLMKVCache * move cache out of args * revert option2 * add get_cache * Revert "add get_cache" This reverts commit a89d9d23810ce594743504fea4bc5cd49e8d4192. * Revert "revert option2" This reverts commit d2b124c1d30717baf482eb887ba5ab3cb09ac51d. * add comments * update comment * Dummy commit for triggering CI * Dummy commit for triggering CI
* refine PatchVLLMKVCache * move cache out of args * revert option2 * add get_cache * Revert "add get_cache" This reverts commit a89d9d23810ce594743504fea4bc5cd49e8d4192. * Revert "revert option2" This reverts commit d2b124c1d30717baf482eb887ba5ab3cb09ac51d. * add comments * update comment * Dummy commit for triggering CI * Dummy commit for triggering CI
* refine PatchVLLMKVCache * move cache out of args * revert option2 * add get_cache * Revert "add get_cache" This reverts commit a89d9d23810ce594743504fea4bc5cd49e8d4192. * Revert "revert option2" This reverts commit d2b124c1d30717baf482eb887ba5ab3cb09ac51d. * add comments * update comment * Dummy commit for triggering CI * Dummy commit for triggering CI
Type of Change
optimization_levelto select the tuning strategyDescription
detail description
JIRA ticket: ILITV-2358
Expected Behavior & Potential Risk
- Start with fp32 model, convert op(s) into lower-precision(
bf16/fp16/int8).- Convert ops into lower precision by op-type wise.
- Convert op into lower precision one by one based on the result of the previous step.
-
0: using conservative strategy-
1: using basic strategy by defaultHow has this PR been tested?
Dependency Change?
None