Skip to content
This repository was archived by the owner on Jun 4, 2025. It is now read-only.

Conversation

@spacemanidol
Copy link

This pull request integrates previous distill QQP and MNLI sparse work with the Sparseml fork of HF.

@spacemanidol spacemanidol requested a review from natuan August 9, 2021 18:19
@spacemanidol
Copy link
Author

@natuan refactored to match your comments

@markurtz markurtz merged commit a0fbe3b into master Oct 8, 2021
KSGulin pushed a commit that referenced this pull request Mar 9, 2022
* Setting up Text Classification for SparseML

* Move teacher's logic out of compute_loss for GLUE

Co-authored-by: Tuan Nguyen <[email protected]>
@dbogunowicz dbogunowicz deleted the BERT-GLUE branch December 5, 2023 10:25
bfineran pushed a commit that referenced this pull request Jun 5, 2024
* Cohere Model Release (#1)

Cohere Model Release

* Remove unnecessary files and code (#2)

Some cleanup

* Delete cohere-model directory (#3)

* Make Fix (#5)

* Pr fixes (#6)

* fixes for pr

* pr fixes for the format

* pr fixes for the format

* src/transformers/models/auto/tokenization_auto.py

* Tokenizer test (#8)

* tokenizer test

* format fix

* Adding Docs and other minor changes (#7)

* Add modeling tests (#9)

* Smol Fix (#11)

* tokenization tests are fixed

* format fixes

* fix pr doc tests

* fix pr doc tests

* fix pr doc tests

* fix pr style check

* small changes in cohere.md

* FIX: Address final comments for transformers integration (#13)

* fix modeling final nits and add proper test file

* for now leave empty tests

* add integration test

* push new test

* fix modeling cohere (#14)

* Update chat templates to use the new API (#15)

---------

Co-authored-by: ahmetustun <[email protected]>
Co-authored-by: Younes Belkada <[email protected]>
Co-authored-by: Matt <[email protected]>
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants