-
Notifications
You must be signed in to change notification settings - Fork 5.3k
[NPU] optimize glm4.7 #19246
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[NPU] optimize glm4.7 #19246
Changes from all commits
8ff90fd
81867d6
9862d71
8aa7d94
0b65526
3d3ddeb
1e667df
f7e583f
49f3d72
c8e0e62
4cd7e42
6e99368
0454218
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -58,13 +58,13 @@ def _rmsnorm_forward_oot( | |
| residual: Optional[torch.Tensor] = None, | ||
| post_residual_addition: Optional[torch.Tensor] = None, | ||
| ) -> Union[torch.Tensor, Tuple[torch.Tensor, torch.Tensor]]: | ||
|
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. The |
||
| from sgl_kernel_npu.norm.add_rmsnorm_bias import add_rmsnorm_bias | ||
|
|
||
| if not x.is_contiguous(): | ||
| x = x.contiguous() | ||
| if residual is not None: | ||
| if post_residual_addition is not None: | ||
| residual = residual + post_residual_addition | ||
| from sgl_kernel_npu.norm.add_rmsnorm_bias import add_rmsnorm_bias | ||
|
|
||
| out, residual_out = add_rmsnorm_bias( | ||
| x, | ||
| residual, | ||
|
|
||
Uh oh!
There was an error while loading. Please reload this page.