-
Notifications
You must be signed in to change notification settings - Fork 31k
[BART] add bart-large-xsum weights #3422
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Codecov Report
@@ Coverage Diff @@
## master #3422 +/- ##
==========================================
- Coverage 77.80% 77.79% -0.01%
==========================================
Files 100 100
Lines 17051 17051
==========================================
- Hits 13266 13265 -1
- Misses 3785 3786 +1
Continue to review full report at Codecov.
|
src/transformers/modeling_bart.py
Outdated
| base_model_prefix = "model" | ||
|
|
||
| def __init__(self, config: BartConfig): | ||
| def __init__(self, config: BartConfig, base_model=None): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What is base_model used for?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Unused, should be deleted.
torch.hub. Finetuning with fairseq and then converting to huggingface should work. I also cleaned it up a bit(num_beams=6, length_penalty=1., min_length=11, max_length=62Context:
These weights are from bart finetuned on the XSum abstractive summarization challenge, which encourages shorter (more abstractive) summaries. It achieves state of the art.
Discussion: