-
-
Notifications
You must be signed in to change notification settings - Fork 1.2k
update flex attention patching for transformers 4.51 #2501
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
… to be in line with transformers v4.51
|
huggingface/transformers#37285 has landed btw |
b98dbaf to
cdb1606
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't know that this can be removed yet, as the released version in transformers had an edge case that affects most of our docker builds. huggingface/transformers#37399
| if self.cfg.flex_attention: | ||
| self.model_kwargs["attn_implementation"] = "flex_attention" | ||
| self.model_config._attn_implementation = ( # pylint: disable=protected-access | ||
| "flex_attention" | ||
| ) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we need to at least keep these lines?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we still even need this PR?
|
superseded by #2469 |
Bumping flex attention monkeypatch to transformers 4.51