Skip to content

Conversation

@ydshieh
Copy link
Collaborator

@ydshieh ydshieh commented Oct 30, 2023

What does this PR do?

Remove some Kosmos-2 copied from as there is recent change in #27086.

Will update them later to match bart, but for now better not to break main CI.

@ydshieh ydshieh requested a review from amyeroberts October 30, 2023 14:24
encoder_states = encoder_states + (hidden_states,)
if self.gradient_checkpointing and self.training:
layer_outputs = self.gradient_checkpointing_func(
layer_outputs = self._gradient_checkpointing_func(
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This needs to be here to match copied. Follow up is #27149 27148

Copy link
Contributor

@amyeroberts amyeroberts left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the quick fix!

And yes - let's move to the new AttentionMaskConverter in a follow-up up :)

@HuggingFaceDocBuilderDev
Copy link

HuggingFaceDocBuilderDev commented Oct 30, 2023

The documentation is not available anymore as the PR was closed or merged.

@ydshieh ydshieh merged commit 3224c0c into main Oct 30, 2023
@ydshieh ydshieh deleted the fix_check_kosmos2 branch October 30, 2023 15:07
EduardoPach pushed a commit to EduardoPach/transformers that referenced this pull request Nov 19, 2023
* fix

* fix

* fix

* fix

---------

Co-authored-by: ydshieh <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants