diff --git a/models/transformers/albert.py b/models/transformers/albert.py index 4bf19b95..2b2b460f 100644 --- a/models/transformers/albert.py +++ b/models/transformers/albert.py @@ -1,9 +1,10 @@ - # labels: name::albert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import AlbertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/auto.py b/models/transformers/auto.py index 4eb6f648..b2bd6a04 100644 --- a/models/transformers/auto.py +++ b/models/transformers/auto.py @@ -1,9 +1,10 @@ - # labels: name::auto author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import AutoModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/bart.py b/models/transformers/bart.py index 8d9894b2..7df877f7 100644 --- a/models/transformers/bart.py +++ b/models/transformers/bart.py @@ -1,9 +1,10 @@ - # labels: name::bart author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BartModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/bert.py b/models/transformers/bert.py index 2bf9e139..2d37da1f 100644 --- a/models/transformers/bert.py +++ b/models/transformers/bert.py @@ -1,9 +1,10 @@ - # labels: name::bert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/bertlmhead.py b/models/transformers/bertlmhead.py index 6fa0cdf6..e687ed6a 100644 --- a/models/transformers/bertlmhead.py +++ b/models/transformers/bertlmhead.py @@ -1,9 +1,10 @@ - # labels: name::bertlmhead author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BertLMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/bigbird.py b/models/transformers/bigbird.py index 164d65a8..59e29205 100644 --- a/models/transformers/bigbird.py +++ b/models/transformers/bigbird.py @@ -1,9 +1,10 @@ - # labels: name::bigbird author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BigBirdModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/bigbirdpegasus.py b/models/transformers/bigbirdpegasus.py index 99fdf2c4..76df275a 100644 --- a/models/transformers/bigbirdpegasus.py +++ b/models/transformers/bigbirdpegasus.py @@ -1,9 +1,10 @@ - # labels: name::bigbirdpegasus author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BigBirdPegasusModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/biogpt.py b/models/transformers/biogpt.py index 07aa30be..2fed5aaa 100644 --- a/models/transformers/biogpt.py +++ b/models/transformers/biogpt.py @@ -1,9 +1,10 @@ - # labels: name::biogpt author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BioGptModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/blenderbot.py b/models/transformers/blenderbot.py index 04b1dfe0..cfccc34c 100644 --- a/models/transformers/blenderbot.py +++ b/models/transformers/blenderbot.py @@ -1,9 +1,10 @@ - # labels: name::blenderbot author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BlenderbotModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/blenderbotsmall.py b/models/transformers/blenderbotsmall.py index 7046b531..8da3f31c 100644 --- a/models/transformers/blenderbotsmall.py +++ b/models/transformers/blenderbotsmall.py @@ -1,9 +1,10 @@ - # labels: name::blenderbotsmall author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BlenderbotSmallModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/bloom.py b/models/transformers/bloom.py index 0cb961fc..5c7919cc 100644 --- a/models/transformers/bloom.py +++ b/models/transformers/bloom.py @@ -1,9 +1,10 @@ - # labels: name::bloom author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import BloomModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -25,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/camembert.py b/models/transformers/camembert.py index 4e139b66..06fcc599 100644 --- a/models/transformers/camembert.py +++ b/models/transformers/camembert.py @@ -1,9 +1,10 @@ - # labels: name::camembert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import CamembertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/canine.py b/models/transformers/canine.py index 32e76ea6..b8b6f464 100644 --- a/models/transformers/canine.py +++ b/models/transformers/canine.py @@ -1,9 +1,10 @@ - # labels: name::canine author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import CanineModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/codegen.py b/models/transformers/codegen.py index 73452a2c..53422c4d 100644 --- a/models/transformers/codegen.py +++ b/models/transformers/codegen.py @@ -1,9 +1,10 @@ - # labels: name::codegen author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import CodeGenModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/convbert.py b/models/transformers/convbert.py index 30c081ee..9645bc09 100644 --- a/models/transformers/convbert.py +++ b/models/transformers/convbert.py @@ -1,9 +1,10 @@ - # labels: name::convbert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import ConvBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/ctrl.py b/models/transformers/ctrl.py index 3fca7811..ea1d7281 100644 --- a/models/transformers/ctrl.py +++ b/models/transformers/ctrl.py @@ -1,9 +1,10 @@ - # labels: name::ctrl author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import CTRLModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/ctrllmhead.py b/models/transformers/ctrllmhead.py index f8944d7f..7b6f9e89 100644 --- a/models/transformers/ctrllmhead.py +++ b/models/transformers/ctrllmhead.py @@ -1,9 +1,10 @@ - # labels: name::ctrllmhead author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import CTRLLMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/deberta.py b/models/transformers/deberta.py index 0befd6c7..73cdc2ce 100644 --- a/models/transformers/deberta.py +++ b/models/transformers/deberta.py @@ -1,9 +1,10 @@ - # labels: name::deberta author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import DebertaModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/debertav2.py b/models/transformers/debertav2.py index 6be84530..e8fa6f17 100644 --- a/models/transformers/debertav2.py +++ b/models/transformers/debertav2.py @@ -1,9 +1,10 @@ - # labels: name::debertav2 author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import DebertaV2Model, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/distilbert.py b/models/transformers/distilbert.py index 89dda024..5fce8649 100644 --- a/models/transformers/distilbert.py +++ b/models/transformers/distilbert.py @@ -1,9 +1,10 @@ - # labels: name::distilbert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import DistilBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/electra.py b/models/transformers/electra.py index efa59eb9..ef7a5d59 100644 --- a/models/transformers/electra.py +++ b/models/transformers/electra.py @@ -1,9 +1,10 @@ - # labels: name::electra author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import ElectraModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/encoderdecoder.py b/models/transformers/encoderdecoder.py index 501014ac..3745b588 100644 --- a/models/transformers/encoderdecoder.py +++ b/models/transformers/encoderdecoder.py @@ -1,9 +1,10 @@ - # labels: name::encoderdecoder author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import EncoderDecoderModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -11,7 +12,9 @@ # Model and input configurations if pretrained: - model = EncoderDecoderModel.from_pretrained("patrickvonplaten/bert2bert_cnn_daily_mail") + model = EncoderDecoderModel.from_pretrained( + "patrickvonplaten/bert2bert_cnn_daily_mail" + ) else: config = AutoConfig.from_pretrained("patrickvonplaten/bert2bert_cnn_daily_mail") model = EncoderDecoderModel(config) @@ -26,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/ernie.py b/models/transformers/ernie.py index a8a323b5..3b06d78a 100644 --- a/models/transformers/ernie.py +++ b/models/transformers/ernie.py @@ -1,9 +1,10 @@ - # labels: name::ernie author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import ErnieModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/erniem.py b/models/transformers/erniem.py index 2975765d..315c76e6 100644 --- a/models/transformers/erniem.py +++ b/models/transformers/erniem.py @@ -1,9 +1,10 @@ - # labels: name::erniem author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import ErnieMModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/esm.py b/models/transformers/esm.py index d28be767..52de7b82 100644 --- a/models/transformers/esm.py +++ b/models/transformers/esm.py @@ -1,9 +1,10 @@ - # labels: name::esm author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import EsmModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/falcon.py b/models/transformers/falcon.py index 028ab7a4..b3c93355 100644 --- a/models/transformers/falcon.py +++ b/models/transformers/falcon.py @@ -1,9 +1,10 @@ - # labels: name::falcon author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import FalconModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/flaubert.py b/models/transformers/flaubert.py index a178b157..8ad758f8 100644 --- a/models/transformers/flaubert.py +++ b/models/transformers/flaubert.py @@ -1,9 +1,10 @@ - # labels: name::flaubert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import FlaubertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/flaubertwithlmhead.py b/models/transformers/flaubertwithlmhead.py index c88483fe..57f8764d 100644 --- a/models/transformers/flaubertwithlmhead.py +++ b/models/transformers/flaubertwithlmhead.py @@ -1,9 +1,10 @@ - # labels: name::flaubertwithlmhead author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import FlaubertWithLMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/fnet.py b/models/transformers/fnet.py index 66e0f86f..f565ddac 100644 --- a/models/transformers/fnet.py +++ b/models/transformers/fnet.py @@ -1,9 +1,10 @@ - # labels: name::fnet author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import FNetModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -27,4 +28,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/funnel.py b/models/transformers/funnel.py index 781e44a0..27fd5a2a 100644 --- a/models/transformers/funnel.py +++ b/models/transformers/funnel.py @@ -1,9 +1,10 @@ - # labels: name::funnel author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import FunnelModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/funnelbase.py b/models/transformers/funnelbase.py index 7ac56cb9..bc796037 100644 --- a/models/transformers/funnelbase.py +++ b/models/transformers/funnelbase.py @@ -1,9 +1,10 @@ - # labels: name::funnelbase author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import FunnelBaseModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/gpt2.py b/models/transformers/gpt2.py index f2b366a5..6f5daa04 100644 --- a/models/transformers/gpt2.py +++ b/models/transformers/gpt2.py @@ -1,9 +1,10 @@ - # labels: name::gpt2 author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import GPT2Model, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/gpt2doubleheads.py b/models/transformers/gpt2doubleheads.py index 57b1d74e..5be70d6f 100644 --- a/models/transformers/gpt2doubleheads.py +++ b/models/transformers/gpt2doubleheads.py @@ -1,9 +1,10 @@ - # labels: name::gpt2doubleheads author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import GPT2DoubleHeadsModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/gpt2lmhead.py b/models/transformers/gpt2lmhead.py index bd14f5ee..57cf3c4f 100644 --- a/models/transformers/gpt2lmhead.py +++ b/models/transformers/gpt2lmhead.py @@ -1,9 +1,10 @@ - # labels: name::gpt2lmhead author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import GPT2LMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/gptbigcode.py b/models/transformers/gptbigcode.py index 8bd3d976..ebec3bc5 100644 --- a/models/transformers/gptbigcode.py +++ b/models/transformers/gptbigcode.py @@ -1,9 +1,10 @@ - # labels: name::gptbigcode author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import GPTBigCodeModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/gptj.py b/models/transformers/gptj.py index 3e170959..c8efb96f 100644 --- a/models/transformers/gptj.py +++ b/models/transformers/gptj.py @@ -1,9 +1,10 @@ - # labels: name::gptj author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import GPTJModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/gptneo.py b/models/transformers/gptneo.py index 275e5e9b..f539b044 100644 --- a/models/transformers/gptneo.py +++ b/models/transformers/gptneo.py @@ -1,9 +1,10 @@ - # labels: name::gptneo author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import GPTNeoModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/gptneox.py b/models/transformers/gptneox.py index 61ac31d0..c2bfef8d 100644 --- a/models/transformers/gptneox.py +++ b/models/transformers/gptneox.py @@ -1,9 +1,10 @@ - # labels: name::gptneox author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import GPTNeoXModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -11,9 +12,13 @@ # Model and input configurations if pretrained: - model = GPTNeoXModel.from_pretrained("trl-internal-testing/tiny-random-GPTNeoXForCausalLM") + model = GPTNeoXModel.from_pretrained( + "trl-internal-testing/tiny-random-GPTNeoXForCausalLM" + ) else: - config = AutoConfig.from_pretrained("trl-internal-testing/tiny-random-GPTNeoXForCausalLM") + config = AutoConfig.from_pretrained( + "trl-internal-testing/tiny-random-GPTNeoXForCausalLM" + ) model = GPTNeoXModel(config) # Make sure the user's sequence length fits within the model's maximum @@ -28,4 +33,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/gptneoxjapanese.py b/models/transformers/gptneoxjapanese.py index 30d000c0..6b76a52f 100644 --- a/models/transformers/gptneoxjapanese.py +++ b/models/transformers/gptneoxjapanese.py @@ -1,9 +1,10 @@ - # labels: name::gptneoxjapanese author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import GPTNeoXJapaneseModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/ibert.py b/models/transformers/ibert.py index 479927d4..cad6a450 100644 --- a/models/transformers/ibert.py +++ b/models/transformers/ibert.py @@ -1,9 +1,10 @@ - # labels: name::ibert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import IBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/led.py b/models/transformers/led.py index ca57e87b..4349eaea 100644 --- a/models/transformers/led.py +++ b/models/transformers/led.py @@ -1,9 +1,10 @@ - # labels: name::led author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import LEDModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -25,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/longformer.py b/models/transformers/longformer.py index 84f4cc59..240e4c97 100644 --- a/models/transformers/longformer.py +++ b/models/transformers/longformer.py @@ -1,9 +1,10 @@ - # labels: name::longformer author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import LongformerModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/longt5.py b/models/transformers/longt5.py index e244f002..e4342f25 100644 --- a/models/transformers/longt5.py +++ b/models/transformers/longt5.py @@ -1,9 +1,10 @@ - # labels: name::longt5 author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import LongT5Model, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -26,4 +27,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/longt5encoder.py b/models/transformers/longt5encoder.py index 4276bc04..fa4fae9b 100644 --- a/models/transformers/longt5encoder.py +++ b/models/transformers/longt5encoder.py @@ -1,9 +1,10 @@ - # labels: name::longt5encoder author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import LongT5EncoderModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -25,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/luke.py b/models/transformers/luke.py index c4a7db01..a832d8ff 100644 --- a/models/transformers/luke.py +++ b/models/transformers/luke.py @@ -1,9 +1,10 @@ - # labels: name::luke author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import LukeModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/m2m100.py b/models/transformers/m2m100.py index 8fecadb6..cb5c3592 100644 --- a/models/transformers/m2m100.py +++ b/models/transformers/m2m100.py @@ -1,9 +1,10 @@ - # labels: name::m2m100 author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import M2M100Model, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/marian.py b/models/transformers/marian.py index f1e12072..4701a225 100644 --- a/models/transformers/marian.py +++ b/models/transformers/marian.py @@ -1,9 +1,10 @@ - # labels: name::marian author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MarianModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/markuplm.py b/models/transformers/markuplm.py index 2f3677a2..cacb5cb5 100644 --- a/models/transformers/markuplm.py +++ b/models/transformers/markuplm.py @@ -1,9 +1,10 @@ - # labels: name::markuplm author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MarkupLMModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/mega.py b/models/transformers/mega.py index 3c855cc0..05a05c50 100644 --- a/models/transformers/mega.py +++ b/models/transformers/mega.py @@ -1,9 +1,10 @@ - # labels: name::mega author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MegaModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -17,7 +18,6 @@ model = MegaModel(config) - inputs = { "input_ids": torch.ones(batch_size, max_seq_length, dtype=torch.long), "attention_mask": torch.ones(batch_size, max_seq_length, dtype=torch.float), @@ -26,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/mobilebert.py b/models/transformers/mobilebert.py index e84d63dd..fd32328b 100644 --- a/models/transformers/mobilebert.py +++ b/models/transformers/mobilebert.py @@ -1,9 +1,10 @@ - # labels: name::mobilebert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MobileBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/mpnet.py b/models/transformers/mpnet.py index 5334789d..ea740b62 100644 --- a/models/transformers/mpnet.py +++ b/models/transformers/mpnet.py @@ -1,9 +1,10 @@ - # labels: name::mpnet author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MPNetModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/mra.py b/models/transformers/mra.py index 946677cf..fbabe1fc 100644 --- a/models/transformers/mra.py +++ b/models/transformers/mra.py @@ -1,9 +1,10 @@ - # labels: name::mra author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MraModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/mt5.py b/models/transformers/mt5.py index 7ac27cec..e99d8dd6 100644 --- a/models/transformers/mt5.py +++ b/models/transformers/mt5.py @@ -1,9 +1,10 @@ - # labels: name::mt5 author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MT5Model, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -26,4 +27,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/mt5encoder.py b/models/transformers/mt5encoder.py index beef4792..124545a2 100644 --- a/models/transformers/mt5encoder.py +++ b/models/transformers/mt5encoder.py @@ -1,9 +1,10 @@ - # labels: name::mt5encoder author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MT5EncoderModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -25,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/mvp.py b/models/transformers/mvp.py index 8350b9ac..92255c70 100644 --- a/models/transformers/mvp.py +++ b/models/transformers/mvp.py @@ -1,9 +1,10 @@ - # labels: name::mvp author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import MvpModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/nezha.py b/models/transformers/nezha.py index f05e46c4..140f3d34 100644 --- a/models/transformers/nezha.py +++ b/models/transformers/nezha.py @@ -1,9 +1,10 @@ - # labels: name::nezha author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import NezhaModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/nystromformer.py b/models/transformers/nystromformer.py index da0f41ac..0cda8095 100644 --- a/models/transformers/nystromformer.py +++ b/models/transformers/nystromformer.py @@ -1,9 +1,10 @@ - # labels: name::nystromformer author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import NystromformerModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/openaigpt.py b/models/transformers/openaigpt.py index 10a741db..bd34ee15 100644 --- a/models/transformers/openaigpt.py +++ b/models/transformers/openaigpt.py @@ -1,9 +1,10 @@ - # labels: name::openaigpt author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import OpenAIGPTModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/openaigptdoubleheads.py b/models/transformers/openaigptdoubleheads.py index fa70a573..2a1ebec2 100644 --- a/models/transformers/openaigptdoubleheads.py +++ b/models/transformers/openaigptdoubleheads.py @@ -1,9 +1,10 @@ - # labels: name::openaigptdoubleheads author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import OpenAIGPTDoubleHeadsModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/openaigptlmhead.py b/models/transformers/openaigptlmhead.py index e4d96826..33703a3f 100644 --- a/models/transformers/openaigptlmhead.py +++ b/models/transformers/openaigptlmhead.py @@ -1,9 +1,10 @@ - # labels: name::openaigptlmhead author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import OpenAIGPTLMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/opt.py b/models/transformers/opt.py index 731632c2..ed5497d4 100644 --- a/models/transformers/opt.py +++ b/models/transformers/opt.py @@ -1,9 +1,10 @@ - # labels: name::opt author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import OPTModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/pegasus.py b/models/transformers/pegasus.py index 2172098e..22c317ae 100644 --- a/models/transformers/pegasus.py +++ b/models/transformers/pegasus.py @@ -1,9 +1,10 @@ - # labels: name::pegasus author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import PegasusModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/prophetnet.py b/models/transformers/prophetnet.py index 90a52c87..60c4da53 100644 --- a/models/transformers/prophetnet.py +++ b/models/transformers/prophetnet.py @@ -1,9 +1,10 @@ - # labels: name::prophetnet author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import ProphetNetModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/reformer.py b/models/transformers/reformer.py index b4ab491a..e2200988 100644 --- a/models/transformers/reformer.py +++ b/models/transformers/reformer.py @@ -1,9 +1,10 @@ - # labels: name::reformer author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import ReformerModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/rembert.py b/models/transformers/rembert.py index 454dc8ff..58c1e0e2 100644 --- a/models/transformers/rembert.py +++ b/models/transformers/rembert.py @@ -1,9 +1,10 @@ - # labels: name::rembert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import RemBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/roberta.py b/models/transformers/roberta.py index d619adb6..335a91a5 100644 --- a/models/transformers/roberta.py +++ b/models/transformers/roberta.py @@ -1,9 +1,10 @@ - # labels: name::roberta author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import RobertaModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/robertaprelayernorm.py b/models/transformers/robertaprelayernorm.py index 9448c573..09f448a7 100644 --- a/models/transformers/robertaprelayernorm.py +++ b/models/transformers/robertaprelayernorm.py @@ -1,9 +1,10 @@ - # labels: name::robertaprelayernorm author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import RobertaPreLayerNormModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -11,7 +12,9 @@ # Model and input configurations if pretrained: - model = RobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40") + model = RobertaPreLayerNormModel.from_pretrained( + "andreasmadsen/efficient_mlm_m0.40" + ) else: config = AutoConfig.from_pretrained("andreasmadsen/efficient_mlm_m0.40") model = RobertaPreLayerNormModel(config) @@ -28,4 +31,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/rocbert.py b/models/transformers/rocbert.py index 63f38ae4..1123709d 100644 --- a/models/transformers/rocbert.py +++ b/models/transformers/rocbert.py @@ -1,9 +1,10 @@ - # labels: name::rocbert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import RoCBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/roformer.py b/models/transformers/roformer.py index 8fabc2d1..61f00d7c 100644 --- a/models/transformers/roformer.py +++ b/models/transformers/roformer.py @@ -1,9 +1,10 @@ - # labels: name::roformer author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import RoFormerModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/fsmt.py b/models/transformers/skip/fsmt.py index b5f80491..86d0ae00 100644 --- a/models/transformers/skip/fsmt.py +++ b/models/transformers/skip/fsmt.py @@ -1,10 +1,11 @@ - # labels: name::fsmt author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Input Error from turnkeyml.parser import parse from transformers import FSMTModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/jukebox.py b/models/transformers/skip/jukebox.py index c73e539f..02d73cbe 100644 --- a/models/transformers/skip/jukebox.py +++ b/models/transformers/skip/jukebox.py @@ -1,10 +1,11 @@ - # labels: name::jukebox author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Input Error from turnkeyml.parser import parse from transformers import JukeboxModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -26,4 +27,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/mbart.py b/models/transformers/skip/mbart.py index 2fab97d2..d5143b92 100644 --- a/models/transformers/skip/mbart.py +++ b/models/transformers/skip/mbart.py @@ -1,10 +1,11 @@ - # labels: name::mbart author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Input Error from turnkeyml.parser import parse from transformers import MBartModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/megatronbert.py b/models/transformers/skip/megatronbert.py index 83ebe025..691c6457 100644 --- a/models/transformers/skip/megatronbert.py +++ b/models/transformers/skip/megatronbert.py @@ -4,6 +4,8 @@ from transformers import MegatronBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/plbart.py b/models/transformers/skip/plbart.py index e2513a77..f4f47880 100644 --- a/models/transformers/skip/plbart.py +++ b/models/transformers/skip/plbart.py @@ -1,10 +1,11 @@ - # labels: name::plbart author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Model not found error from turnkeyml.parser import parse from transformers import PLBartModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/qdqbert.py b/models/transformers/skip/qdqbert.py index 01dab3ed..77461f21 100644 --- a/models/transformers/skip/qdqbert.py +++ b/models/transformers/skip/qdqbert.py @@ -1,10 +1,11 @@ - # labels: name::qdqbert author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Input error from turnkeyml.parser import parse from transformers import QDQBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/qdqbertlmhead.py b/models/transformers/skip/qdqbertlmhead.py index 34667bc6..725fdf69 100644 --- a/models/transformers/skip/qdqbertlmhead.py +++ b/models/transformers/skip/qdqbertlmhead.py @@ -1,10 +1,11 @@ - # labels: name::qdqbertlmhead author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Input error from turnkeyml.parser import parse from transformers import QDQBertLMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/rag.py b/models/transformers/skip/rag.py index 58b1a96b..39ac2c87 100644 --- a/models/transformers/skip/rag.py +++ b/models/transformers/skip/rag.py @@ -1,10 +1,11 @@ - # labels: name::rag author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Input error from turnkeyml.parser import parse from transformers import RagModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -26,4 +27,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/rwkv.py b/models/transformers/skip/rwkv.py similarity index 92% rename from models/transformers/rwkv.py rename to models/transformers/skip/rwkv.py index 95ca39f7..0338968f 100644 --- a/models/transformers/rwkv.py +++ b/models/transformers/skip/rwkv.py @@ -1,9 +1,11 @@ - # labels: name::rwkv author::transformers task::Generative_AI license::apache-2.0 +# Skip reason: takes over 3 hours to export to ONNX from turnkeyml.parser import parse from transformers import RwkvModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/xlmprophetnet.py b/models/transformers/skip/xlmprophetnet.py index f99aa631..dae51447 100644 --- a/models/transformers/skip/xlmprophetnet.py +++ b/models/transformers/skip/xlmprophetnet.py @@ -1,10 +1,11 @@ - # labels: name::xlmprophetnet author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Input error from turnkeyml.parser import parse from transformers import XLMProphetNetModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -12,9 +13,13 @@ # Model and input configurations if pretrained: - model = XLMProphetNetModel.from_pretrained("patrickvonplaten/xprophetnet-large-uncased-standalone") + model = XLMProphetNetModel.from_pretrained( + "patrickvonplaten/xprophetnet-large-uncased-standalone" + ) else: - config = AutoConfig.from_pretrained("patrickvonplaten/xprophetnet-large-uncased-standalone") + config = AutoConfig.from_pretrained( + "patrickvonplaten/xprophetnet-large-uncased-standalone" + ) model = XLMProphetNetModel(config) # Make sure the user's sequence length fits within the model's maximum @@ -30,4 +35,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/skip/xlmrobertaxl.py b/models/transformers/skip/xlmrobertaxl.py index d53ddbbc..5af57057 100644 --- a/models/transformers/skip/xlmrobertaxl.py +++ b/models/transformers/skip/xlmrobertaxl.py @@ -1,10 +1,11 @@ - # labels: name::xlmrobertaxl author::transformers task::Generative_AI license::apache-2.0 # Skip reason: Model not found from turnkeyml.parser import parse from transformers import XLMRobertaXLModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -29,4 +30,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/splinter.py b/models/transformers/splinter.py index 611bdaa6..7ab40ecd 100644 --- a/models/transformers/splinter.py +++ b/models/transformers/splinter.py @@ -1,9 +1,10 @@ - # labels: name::splinter author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import SplinterModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/squeezebert.py b/models/transformers/squeezebert.py index b570de82..34e20425 100644 --- a/models/transformers/squeezebert.py +++ b/models/transformers/squeezebert.py @@ -1,9 +1,10 @@ - # labels: name::squeezebert author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import SqueezeBertModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/switchtransformers.py b/models/transformers/switchtransformers.py index ccfd89e9..b5866291 100644 --- a/models/transformers/switchtransformers.py +++ b/models/transformers/switchtransformers.py @@ -1,9 +1,10 @@ - # labels: name::switchtransformers author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import SwitchTransformersModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -26,4 +27,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/switchtransformersencoder.py b/models/transformers/switchtransformersencoder.py index 37bfb2c0..857c9bf3 100644 --- a/models/transformers/switchtransformersencoder.py +++ b/models/transformers/switchtransformersencoder.py @@ -1,9 +1,10 @@ - # labels: name::switchtransformersencoder author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import SwitchTransformersEncoderModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -25,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/t5.py b/models/transformers/t5.py index 99262650..39933ca1 100644 --- a/models/transformers/t5.py +++ b/models/transformers/t5.py @@ -1,9 +1,10 @@ - # labels: name::t5 author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import T5Model, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -25,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/t5encoder.py b/models/transformers/t5encoder.py index 33a53c2d..8bd01ec8 100644 --- a/models/transformers/t5encoder.py +++ b/models/transformers/t5encoder.py @@ -1,9 +1,10 @@ - # labels: name::t5encoder author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import T5EncoderModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -25,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/transfoxl.py b/models/transformers/transfoxl.py index 549d52e6..f7b89225 100644 --- a/models/transformers/transfoxl.py +++ b/models/transformers/transfoxl.py @@ -1,9 +1,10 @@ - # labels: name::transfoxl author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import TransfoXLModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -27,4 +28,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/transfoxllmhead.py b/models/transformers/transfoxllmhead.py index 075e4c95..33a837e9 100644 --- a/models/transformers/transfoxllmhead.py +++ b/models/transformers/transfoxllmhead.py @@ -1,9 +1,10 @@ - # labels: name::transfoxllmhead author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import TransfoXLLMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -24,4 +25,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/umt5.py b/models/transformers/umt5.py index 803bca22..3ed01d3e 100644 --- a/models/transformers/umt5.py +++ b/models/transformers/umt5.py @@ -1,9 +1,10 @@ - # labels: name::umt5 author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import UMT5Model, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -26,4 +27,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/umt5encoder.py b/models/transformers/umt5encoder.py index eedbe5d8..2e3fe228 100644 --- a/models/transformers/umt5encoder.py +++ b/models/transformers/umt5encoder.py @@ -1,9 +1,10 @@ - # labels: name::umt5encoder author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import UMT5EncoderModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -25,4 +26,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/xglm.py b/models/transformers/xglm.py index 486141f1..a05be3b5 100644 --- a/models/transformers/xglm.py +++ b/models/transformers/xglm.py @@ -1,9 +1,10 @@ - # labels: name::xglm author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import XGLMModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/xlm.py b/models/transformers/xlm.py index bc1db168..acb053e7 100644 --- a/models/transformers/xlm.py +++ b/models/transformers/xlm.py @@ -1,9 +1,10 @@ - # labels: name::xlm author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import XLMModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/xlmroberta.py b/models/transformers/xlmroberta.py index 04d1d7d8..8308d428 100644 --- a/models/transformers/xlmroberta.py +++ b/models/transformers/xlmroberta.py @@ -1,9 +1,10 @@ - # labels: name::xlmroberta author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import XLMRobertaModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/xlmwithlmhead.py b/models/transformers/xlmwithlmhead.py index 7cf2b461..83eb3774 100644 --- a/models/transformers/xlmwithlmhead.py +++ b/models/transformers/xlmwithlmhead.py @@ -1,9 +1,10 @@ - # labels: name::xlmwithlmhead author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import XLMWithLMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/xlnet.py b/models/transformers/xlnet.py index e7fe3b3d..4276ecb6 100644 --- a/models/transformers/xlnet.py +++ b/models/transformers/xlnet.py @@ -1,9 +1,10 @@ - # labels: name::xlnet author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import XLNetModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -24,4 +25,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/xlnetlmhead.py b/models/transformers/xlnetlmhead.py index 71a23c2e..820f475e 100644 --- a/models/transformers/xlnetlmhead.py +++ b/models/transformers/xlnetlmhead.py @@ -1,9 +1,10 @@ - # labels: name::xlnetlmhead author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import XLNetLMHeadModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -24,4 +25,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/xmod.py b/models/transformers/xmod.py index 46ef57c8..87e8f93a 100644 --- a/models/transformers/xmod.py +++ b/models/transformers/xmod.py @@ -1,9 +1,10 @@ - # labels: name::xmod author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import XmodModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -26,4 +27,3 @@ # Call model model(**inputs) - \ No newline at end of file diff --git a/models/transformers/yoso.py b/models/transformers/yoso.py index fb495f8a..78a5f4e8 100644 --- a/models/transformers/yoso.py +++ b/models/transformers/yoso.py @@ -1,9 +1,10 @@ - # labels: name::yoso author::transformers task::Generative_AI license::apache-2.0 from turnkeyml.parser import parse from transformers import YosoModel, AutoConfig import torch +torch.manual_seed(0) + # Parsing command-line arguments pretrained, batch_size, max_seq_length = parse( ["pretrained", "batch_size", "max_seq_length"] @@ -28,4 +29,3 @@ # Call model model(**inputs) - \ No newline at end of file