Skip to content

Conversation

@matthewdouglas
Copy link
Member

What does this PR do?

Fixes a test failure that resulted from the introduction of OPTSdpaAttention as the default implementation in #33298. Discussed internally on Slack.

Before submitting

  • This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • Did you read the contributor guideline,
    Pull Request section?
  • Was this discussed/approved via a Github issue or the forum? Please add a link
    to it if that's the case.
  • Did you make sure to update the documentation with your changes? Here are the
    documentation guidelines, and
    here are tips on formatting docstrings.
  • Did you write any new necessary tests?

Who can review?

@BenjaminBossan @SunMarc

@matthewdouglas matthewdouglas added the Tests Related to tests label Oct 25, 2024
Copy link
Member

@BenjaminBossan BenjaminBossan left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for identifying the issue and providing a fix. I tested the change locally and the tests now pass for me.

@matthewdouglas matthewdouglas changed the title Fix te Fix bnb training test failure Oct 25, 2024
@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

@matthewdouglas matthewdouglas merged commit e447185 into main Oct 25, 2024
13 checks passed
@matthewdouglas matthewdouglas deleted the fix-test-bnb-opt-sdpa branch October 25, 2024 14:23
@SunMarc
Copy link
Member

SunMarc commented Oct 25, 2024

I saw that the issue was with the OPTSdpaAttention but why it doesn't work with it ? Any ideas ? cc @matthewdouglas

@BenjaminBossan
Copy link
Member

The issue was that the old check "OPTAttention" in repr(type(module)) didn't match any layer anymore, hence no LoRA layers were added, hence no trainable parameters were found, resulting in the error.

@SunMarc
Copy link
Member

SunMarc commented Oct 25, 2024

Oh, I read the PR too fast. Thanks ! It makes sense now

BernardZach pushed a commit to BernardZach/transformers that referenced this pull request Dec 5, 2024
* Fix bnb training test: compatibility with OPTSdpaAttention
ylacombe added a commit that referenced this pull request Dec 10, 2024
* Support BatchNorm in Hubert pos_conv_emb as in fairseq

* Correct the new defaults (#34377)

* Correct the new defaults

* CIs

* add check

* Update utils.py

* Update utils.py

* Add the max_length in generate test checking shape without passing length

* style

* CIs

* fix fx CI issue

* [auto. ping] Avoid sending empty info + add more team members (#34383)

* update

* update

---------

Co-authored-by: ydshieh <[email protected]>

* Fix glm  (#34388)

* Fix duplicated

* fix import

* Use non nested images and batched text Idefics2/3  (#34222)

* add support for non nested images and add tests

* add tests error scenario

* fix style

* added single and no image to error tests

* Fix onnx non-expotable inplace aten op (#34376)

* fix onnx non-expotable inplace op

* mistral, qwen2, qwen2_vl, starcoder2

* fixup copies

* Fix right padding in LLaVA models (#34305)

* fix right pad llavas

* device mismatch

* no filter (#34391)

* no filter

* no filter

* no filter

---------

Co-authored-by: ydshieh <[email protected]>

* SynthID: better example (#34372)

* better example

* Update src/transformers/generation/configuration_utils.py

* Update src/transformers/generation/logits_process.py

* nits

* Tests: upgrade `test_eager_matches_sdpa_generate` (#34386)

* Fix bnb training test failure (#34414)

* Fix bnb training test: compatibility with OPTSdpaAttention

* Avoid check expected exception when it is on CUDA (#34408)

* update

* update

---------

Co-authored-by: ydshieh <[email protected]>

* Fix typos in agents_advanced.md (#34405)

* [docs] Cache implementations (#34325)

cache

* [run-slow] hubert

* Support BatchNorm in Hubert pos_conv_emb as in fairseq
Add conversion integration test, and make batchnorm explicit variable

* Support BatchNorm in Hubert pos_conv_emb as in fairseq
fix make fixup styling changes

* [run-slow] hubert

* Support BatchNorm in Hubert pos_conv_emb as in fairseq

* [run-slow] hubert

* Support BatchNorm in Hubert pos_conv_emb as in fairseq
Add conversion integration test, and make batchnorm explicit variable

* Support BatchNorm in Hubert pos_conv_emb as in fairseq
fix make fixup styling changes

* [run-slow] hubert

* [run-slow] hubert

---------

Co-authored-by: Cyril Vallez <[email protected]>
Co-authored-by: Yih-Dar <[email protected]>
Co-authored-by: ydshieh <[email protected]>
Co-authored-by: Yoni Gozlan <[email protected]>
Co-authored-by: Ilyas Moutawwakil <[email protected]>
Co-authored-by: Raushan Turganbay <[email protected]>
Co-authored-by: Joao Gante <[email protected]>
Co-authored-by: Matthew Douglas <[email protected]>
Co-authored-by: Rudy Delouya <[email protected]>
Co-authored-by: Steven Liu <[email protected]>
Co-authored-by: Yoach Lacombe <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Tests Related to tests

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants