diff --git a/src/transformers/models/swin/modeling_swin.py b/src/transformers/models/swin/modeling_swin.py index 2cf1d33a5113..45a7aa718cf0 100644 --- a/src/transformers/models/swin/modeling_swin.py +++ b/src/transformers/models/swin/modeling_swin.py @@ -942,6 +942,12 @@ def _set_gradient_checkpointing(self, module, value=False): @add_start_docstrings( "The bare Swin Model transformer outputting raw hidden-states without any specific head on top.", SWIN_START_DOCSTRING, + """ + add_pooling_layer (`bool`, *optional*, defaults to `True`): + Whether or not to apply pooling layer. + use_mask_token (`bool`, *optional*, defaults to `False`): + Whether or not to create and apply mask tokens in the embedding layer. + """, ) class SwinModel(SwinPreTrainedModel): def __init__(self, config, add_pooling_layer=True, use_mask_token=False): diff --git a/utils/check_docstrings.py b/utils/check_docstrings.py index 650b519eaa57..748137fe4af5 100644 --- a/utils/check_docstrings.py +++ b/utils/check_docstrings.py @@ -508,7 +508,6 @@ "SqueezeBertTokenizerFast", "SummarizationPipeline", "Swin2SRImageProcessor", - "SwinModel", "Swinv2Model", "SwitchTransformersConfig", "T5Config",