From f119d0b9181aa8a3068b5b7195fbaf0e5218fd5c Mon Sep 17 00:00:00 2001 From: MurugeshMarvel Date: Sat, 17 Jul 2021 20:24:55 +0530 Subject: [PATCH 1/2] changes to support transformers v4.x --- models/tabformer_bert.py | 6 +++--- models/tabformer_gpt2.py | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/models/tabformer_bert.py b/models/tabformer_bert.py index bc7c0ca..f0748e4 100644 --- a/models/tabformer_bert.py +++ b/models/tabformer_bert.py @@ -2,8 +2,8 @@ from torch import nn from torch.nn import CrossEntropyLoss -from transformers.modeling_bert import ACT2FN, BertLayerNorm -from transformers.modeling_bert import BertForMaskedLM +from transformers.models.bert.modeling_bert import ACT2FN +from transformers.models.bert.modeling_bert import BertForMaskedLM from transformers.configuration_bert import BertConfig from models.custom_criterion import CustomAdaptiveLogSoftmax @@ -37,7 +37,7 @@ def __init__(self, config): self.transform_act_fn = ACT2FN[config.hidden_act] else: self.transform_act_fn = config.hidden_act - self.LayerNorm = BertLayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) def forward(self, hidden_states): hidden_states = self.dense(hidden_states) diff --git a/models/tabformer_gpt2.py b/models/tabformer_gpt2.py index ef814cb..66665bc 100644 --- a/models/tabformer_gpt2.py +++ b/models/tabformer_gpt2.py @@ -1,6 +1,6 @@ from torch.nn import CrossEntropyLoss -from transformers.modeling_gpt2 import GPT2LMHeadModel +from transformers.models.gpt2.modeling_gpt2 import GPT2LMHeadModel class TabFormerGPT2LMHeadModel(GPT2LMHeadModel): From a0ec2b3789a813c077889a3bdc003e683f51f40c Mon Sep 17 00:00:00 2001 From: Murugesh Marvel Date: Sat, 17 Jul 2021 20:48:30 +0530 Subject: [PATCH 2/2] changes to support transformers v4.x --- models/tabformer_bert.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/models/tabformer_bert.py b/models/tabformer_bert.py index f0748e4..def5573 100644 --- a/models/tabformer_bert.py +++ b/models/tabformer_bert.py @@ -4,7 +4,7 @@ from transformers.models.bert.modeling_bert import ACT2FN from transformers.models.bert.modeling_bert import BertForMaskedLM -from transformers.configuration_bert import BertConfig +from transformers.models.bert.configuration_bert import BertConfig from models.custom_criterion import CustomAdaptiveLogSoftmax @@ -190,4 +190,4 @@ def forward( sequence_output = outputs[0] # [bsz * seqlen * hidden] - return sequence_output \ No newline at end of file + return sequence_output