You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently, NeMo 25.02rc2 container uses TE 1.14.0+87fbe81, which blocks certain attention use cases for pad_between_seqs=True when qkv_format=thd. Please include this PR in the next container build: NVIDIA/TransformerEngine#1387
Describe the bug
Currently, NeMo 25.02rc2 container uses TE 1.14.0+87fbe81, which blocks certain attention use cases for
pad_between_seqs=True
whenqkv_format=thd
. Please include this PR in the next container build:NVIDIA/TransformerEngine#1387
In particular, these lines are the fix:
https://github.com/NVIDIA/TransformerEngine/blob/f0d22ca12f574233053da20516997e45d99eb65c/transformer_engine/pytorch/attention.py#L7457-L7463
The text was updated successfully, but these errors were encountered: