-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
placeholder from Speechllm selene to main #13
Conversation
approve the nit change
…#2) approve as no need to review
… to guard the llm input
* Merge heh and zhehuai's initial version of frozen am+llm The previous differences are summarized here: https://docs.google.com/document/d/1zNI4hC6vJtUfcHbrUSPaMuYWRBQdN_36H0P2NiBiuPY/edit This PR includes 1. Finish merging the model, dataset, and config code 2. Previous tests are still enabled and passed (prepare_llm_input, training_step, validation_step) 3. the example training script with LS960 has been run to make sure the training pipeline works The major remaining works are listed here https://docs.google.com/document/d/1o0AM7v4gcTQkPZjE0Vl9TTX4vYnGTrbXEFGWh0UhGlk/edit#bookmark=id.pzvdadt5oxyw --------- Co-authored-by: He Huang (Steve) <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
init them from the ckpt Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: stevehuang52 <[email protected]>
…into speechllm_selene_he
Signed-off-by: stevehuang52 <[email protected]>
add tarred datasets
Signed-off-by: zhehuaichen <[email protected]>
…peechllm_selene Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: stevehuang52 <[email protected]>
Signed-off-by: stevehuang52 <[email protected]>
Signed-off-by: stevehuang52 <[email protected]>
fix bucketing dataset
Signed-off-by: zhehuaichen <[email protected]>
multitask understanding; also fix bleu implementation Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
…models Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
Signed-off-by: zhehuaichen <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
left some items to run down
…g update in training Signed-off-by: zhehuaichen <[email protected]>
8c98989
to
3f5a351
Compare
@@ -541,46 +593,88 @@ def inference_epoch_end(self, outputs, mode, data_cfg): | |||
averaged_metric = 0.0 if monitor_mode == 'max' else 1e5 | |||
|
|||
if mode == 'validation': | |||
self.log("validation_loss", averaged_loss) | |||
self.log("validation_loss", averaged_loss, batch_size=1, sync_dist=True) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
may not be necessary, I added batch_size=1, sync_dist=True
just because other places have them. it's ok to keep the original code.
def on_test_epoch_end(self): | ||
self.on_inference_epoch_end(self.cfg.data.test_ds) | ||
return super().on_test_epoch_end() | ||
# def on_test_epoch_end(self): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
commented out since they are not actually called in PTL1.8
What does this PR do ?
Add a one line overview of what this PR aims to accomplish.
Collection: [Note which collection this PR will affect]
Changelog
Usage
# Add a code snippet demonstrating how to use this
Before your PR is "Ready for review"
Pre checks:
PR Type:
If you haven't finished some of the above items you can still open "Draft" PR.
Who can review?
Anyone in the community is free to review the PR once the checks have passed.
Contributor guidelines contains specific people who can review PRs to various areas.
Additional Information