Skip to content

Commit 32eb29f

Browse files
authored
Fix doc examples: modify config before super().__init__ (huggingface#14697)
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
1 parent 48bf7e4 commit 32eb29f

File tree

10 files changed

+10
-10
lines changed

10 files changed

+10
-10
lines changed

src/transformers/models/bart/modeling_bart.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1662,10 +1662,10 @@ def forward(self, *args, **kwargs):
16621662

16631663
class BartForCausalLM(BartPretrainedModel):
16641664
def __init__(self, config):
1665-
super().__init__(config)
16661665
config = copy.deepcopy(config)
16671666
config.is_decoder = True
16681667
config.is_encoder_decoder = False
1668+
super().__init__(config)
16691669
self.model = BartDecoderWrapper(config)
16701670

16711671
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

src/transformers/models/bigbird_pegasus/modeling_bigbird_pegasus.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2865,10 +2865,10 @@ def forward(self, *args, **kwargs):
28652865
# Copied from transformers.models.bart.modeling_bart.BartForCausalLM with Bart->BigBirdPegasus, 'facebook/bart-large'->"google/bigbird-pegasus-large-arxiv"
28662866
class BigBirdPegasusForCausalLM(BigBirdPegasusPreTrainedModel):
28672867
def __init__(self, config):
2868-
super().__init__(config)
28692868
config = copy.deepcopy(config)
28702869
config.is_decoder = True
28712870
config.is_encoder_decoder = False
2871+
super().__init__(config)
28722872
self.model = BigBirdPegasusDecoderWrapper(config)
28732873

28742874
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

src/transformers/models/blenderbot/modeling_blenderbot.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1400,10 +1400,10 @@ def forward(self, *args, **kwargs):
14001400
# Copied from transformers.models.bart.modeling_bart.BartForCausalLM with Bart->Blenderbot
14011401
class BlenderbotForCausalLM(BlenderbotPreTrainedModel):
14021402
def __init__(self, config):
1403-
super().__init__(config)
14041403
config = copy.deepcopy(config)
14051404
config.is_decoder = True
14061405
config.is_encoder_decoder = False
1406+
super().__init__(config)
14071407
self.model = BlenderbotDecoderWrapper(config)
14081408

14091409
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

src/transformers/models/blenderbot_small/modeling_blenderbot_small.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1374,10 +1374,10 @@ def forward(self, *args, **kwargs):
13741374
# Copied from transformers.models.bart.modeling_bart.BartForCausalLM with Bart->BlenderbotSmall
13751375
class BlenderbotSmallForCausalLM(BlenderbotSmallPreTrainedModel):
13761376
def __init__(self, config):
1377-
super().__init__(config)
13781377
config = copy.deepcopy(config)
13791378
config.is_decoder = True
13801379
config.is_encoder_decoder = False
1380+
super().__init__(config)
13811381
self.model = BlenderbotSmallDecoderWrapper(config)
13821382

13831383
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

src/transformers/models/marian/modeling_marian.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1397,10 +1397,10 @@ def forward(self, *args, **kwargs):
13971397
# Copied from transformers.models.bart.modeling_bart.BartForCausalLM with Bart->Marian
13981398
class MarianForCausalLM(MarianPreTrainedModel):
13991399
def __init__(self, config):
1400-
super().__init__(config)
14011400
config = copy.deepcopy(config)
14021401
config.is_decoder = True
14031402
config.is_encoder_decoder = False
1403+
super().__init__(config)
14041404
self.model = MarianDecoderWrapper(config)
14051405

14061406
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

src/transformers/models/mbart/modeling_mbart.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1665,10 +1665,10 @@ def forward(self, *args, **kwargs):
16651665
# Copied from transformers.models.bart.modeling_bart.BartForCausalLM with Bart->MBart
16661666
class MBartForCausalLM(MBartPreTrainedModel):
16671667
def __init__(self, config):
1668-
super().__init__(config)
16691668
config = copy.deepcopy(config)
16701669
config.is_decoder = True
16711670
config.is_encoder_decoder = False
1671+
super().__init__(config)
16721672
self.model = MBartDecoderWrapper(config)
16731673

16741674
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

src/transformers/models/pegasus/modeling_pegasus.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1486,10 +1486,10 @@ def forward(self, *args, **kwargs):
14861486

14871487
class PegasusForCausalLM(PegasusPreTrainedModel):
14881488
def __init__(self, config):
1489-
super().__init__(config)
14901489
config = copy.deepcopy(config)
14911490
config.is_decoder = True
14921491
config.is_encoder_decoder = False
1492+
super().__init__(config)
14931493
self.model = PegasusDecoderWrapper(config)
14941494

14951495
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

src/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -744,10 +744,10 @@ def forward(self, *args, **kwargs):
744744
)
745745
class Speech2Text2ForCausalLM(Speech2Text2PreTrainedModel):
746746
def __init__(self, config):
747-
super().__init__(config)
748747
config = copy.deepcopy(config)
749748
config.is_decoder = True
750749
config.is_encoder_decoder = False
750+
super().__init__(config)
751751
self.model = Speech2Text2DecoderWrapper(config)
752752

753753
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

src/transformers/models/trocr/modeling_trocr.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -777,10 +777,10 @@ def forward(self, *args, **kwargs):
777777
)
778778
class TrOCRForCausalLM(TrOCRPreTrainedModel):
779779
def __init__(self, config):
780-
super().__init__(config)
781780
config = copy.deepcopy(config)
782781
config.is_decoder = True
783782
config.is_encoder_decoder = False
783+
super().__init__(config)
784784
self.model = TrOCRDecoderWrapper(config)
785785

786786
self.output_projection = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

templates/adding_a_new_model/cookiecutter-template-{{cookiecutter.modelname}}/modeling_{{cookiecutter.lowercase_modelname}}.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3173,10 +3173,10 @@ def forward(self, *args, **kwargs):
31733173
# Copied from transformers.models.bart.modeling_bart.BartForCausalLM with Bart->{{cookiecutter.camelcase_modelname}}
31743174
class {{cookiecutter.camelcase_modelname}}ForCausalLM({{cookiecutter.camelcase_modelname}}PreTrainedModel):
31753175
def __init__(self, config):
3176-
super().__init__(config)
31773176
config = copy.deepcopy(config)
31783177
config.is_decoder = True
31793178
config.is_encoder_decoder = False
3179+
super().__init__(config)
31803180
self.model = {{cookiecutter.camelcase_modelname}}DecoderWrapper(config)
31813181

31823182
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)

0 commit comments

Comments
 (0)