Skip to content

Commit 10ddfb0

Browse files
authored
Add more missing arguments (#40354)
Add missing arguments Signed-off-by: cyy <cyyever@outlook.com>
1 parent d10603f commit 10ddfb0

37 files changed

+53
-55
lines changed

src/transformers/models/bamba/modular_bamba.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -708,7 +708,7 @@ class BambaRMSNorm(LlamaRMSNorm):
708708

709709
class BambaDecoderLayer(JambaAttentionDecoderLayer):
710710
def __init__(self, config: BambaConfig, layer_idx: int, layer_type: str = "mamba"):
711-
super().__init__()
711+
super().__init__(config, layer_idx)
712712

713713
del self.self_attn
714714

src/transformers/models/cohere2/modular_cohere2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -271,7 +271,7 @@ class Cohere2LayerNorm(CohereLayerNorm):
271271
pass
272272

273273

274-
class Cohere2Attention(CohereAttention, nn.Module):
274+
class Cohere2Attention(CohereAttention):
275275
"""Multi-headed attention from 'Attention Is All You Need' paper"""
276276

277277
def __init__(self, config: Cohere2Config, layer_idx: Optional[int] = None):

src/transformers/models/d_fine/modular_d_fine.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -898,7 +898,7 @@ def __init__(self, config: DFineConfig):
898898

899899
class DFineForObjectDetection(RTDetrForObjectDetection, DFinePreTrainedModel):
900900
def __init__(self, config: DFineConfig):
901-
DFinePreTrainedModel.__init__(config)
901+
DFinePreTrainedModel.__init__(self, config)
902902

903903
# D-FINE encoder-decoder model
904904
self.eval_idx = config.eval_idx if config.eval_idx >= 0 else config.decoder_layers + config.eval_idx

src/transformers/models/data2vec/modular_data2vec_audio.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -112,7 +112,7 @@ def forward(self, hidden_states):
112112
return hidden_states
113113

114114

115-
class Data2VecAudioFeatureEncoder(Wav2Vec2FeatureEncoder, nn.Module):
115+
class Data2VecAudioFeatureEncoder(Wav2Vec2FeatureEncoder):
116116
def __init__(self, config):
117117
nn.Module.__init__(self)
118118
self.conv_layers = nn.ModuleList(
@@ -183,7 +183,7 @@ def load_adapter(self):
183183

184184
class Data2VecAudioModel(Data2VecAudioPreTrainedModel, Wav2Vec2Model):
185185
def __init__(self, config: Data2VecAudioConfig):
186-
Data2VecAudioPreTrainedModel.__init__(config)
186+
Data2VecAudioPreTrainedModel.__init__(self, config)
187187
self.config = config
188188
self.feature_extractor = Data2VecAudioFeatureEncoder(config)
189189
self.feature_projection = Data2VecAudioFeatureProjection(config)
@@ -215,7 +215,7 @@ def forward(self, **super_kwargs):
215215

216216
class Data2VecAudioForCTC(Data2VecAudioPreTrainedModel, Wav2Vec2ForCTC):
217217
def __init__(self, config):
218-
Data2VecAudioPreTrainedModel.__init__(config)
218+
Data2VecAudioPreTrainedModel.__init__(self, config)
219219

220220
self.data2vec_audio = Data2VecAudioModel(config)
221221
self.dropout = nn.Dropout(config.final_dropout)

src/transformers/models/dia/modular_dia.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -107,7 +107,7 @@ class DiaRotaryEmbedding(LlamaRotaryEmbedding):
107107
pass
108108

109109

110-
class DiaSelfAttention(LlamaAttention, nn.Module):
110+
class DiaSelfAttention(LlamaAttention):
111111
"""Multi-headed attention from 'Attention Is All You Need' paper"""
112112

113113
def __init__(self, config: Union[DiaEncoderConfig, DiaDecoderConfig], layer_idx: int, is_causal: bool = False):

src/transformers/models/dots1/modular_dots1.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,7 @@ class Dots1TopkRouter(DeepseekV3TopkRouter):
6262

6363
class Dots1DecoderLayer(DeepseekV3DecoderLayer):
6464
def __init__(self, config: Dots1Config, layer_idx: int):
65-
super().__init__()
65+
super().__init__(config, layer_idx)
6666
self.attention_type = config.layer_types[layer_idx]
6767

6868

src/transformers/models/ernie4_5/modular_ernie4_5.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ def apply_rotary_pos_emb(q, k, cos, sin, position_ids=None, unsqueeze_dim=1):
8484

8585
class Ernie4_5MLP(LlamaMLP):
8686
def __init__(self, config: Ernie4_5Config):
87-
super().__init__()
87+
super().__init__(config)
8888

8989
self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.use_bias)
9090
self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.use_bias)

src/transformers/models/esm/openfold_utils/residue_constants.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -541,7 +541,6 @@ def make_bond_key(atom1_name: str, atom2_name: str) -> str:
541541

542542
# A compact atom encoding with 14 columns
543543
# pylint: disable=line-too-long
544-
# pylint: disable=bad-whitespace
545544
restype_name_to_atom14_names: dict[str, list[str]] = {
546545
"ALA": ["N", "CA", "C", "O", "CB", "", "", "", "", "", "", "", "", ""],
547546
"ARG": ["N", "CA", "C", "O", "CB", "CG", "CD", "NE", "CZ", "NH1", "NH2", "", "", ""],
@@ -566,7 +565,6 @@ def make_bond_key(atom1_name: str, atom2_name: str) -> str:
566565
"UNK": ["", "", "", "", "", "", "", "", "", "", "", "", "", ""],
567566
}
568567
# pylint: enable=line-too-long
569-
# pylint: enable=bad-whitespace
570568

571569

572570
# This is the standard residue order when coding AA type as a number.

src/transformers/models/evolla/modular_evolla.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@
6565

6666
class EvollaSaProtEmbeddings(EsmEmbeddings):
6767
def __init__(self, config):
68-
super().__init__()
68+
super().__init__(config)
6969
# remove the position_ids in EsmEmbeddings
7070
self.position_ids = None
7171

@@ -127,7 +127,7 @@ def forward(self, q: torch.Tensor, k: torch.Tensor) -> tuple[torch.Tensor, torch
127127
)
128128

129129

130-
class EvollaSaProtSelfAttention(EsmSelfAttention, nn.Module):
130+
class EvollaSaProtSelfAttention(EsmSelfAttention):
131131
def __init__(self, config, position_embedding_type=None, layer_idx=None, is_cross_attention=False):
132132
nn.Module.__init__(self)
133133
self.config = config

src/transformers/models/falcon_h1/modeling_falcon_h1.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1007,7 +1007,7 @@ def forward(
10071007

10081008

10091009
class FalconH1MLP(nn.Module):
1010-
def __init__(self, config: FalconH1Config = None):
1010+
def __init__(self, config: FalconH1Config):
10111011
super().__init__()
10121012
self.config = config
10131013
self.hidden_size = config.hidden_size

0 commit comments

Comments
 (0)