From b1ff0b2ae7d368b7db3a8a8472a29cc195d278d8 Mon Sep 17 00:00:00 2001 From: Andrey Kulagin Date: Fri, 17 Apr 2020 17:33:24 +0300 Subject: [PATCH] Fix bug in examples: double wrap into DataParallel during eval --- examples/hans/test_hans.py | 2 +- examples/mm-imdb/run_mmimdb.py | 2 +- examples/ner/run_ner.py | 2 +- examples/run_language_modeling.py | 2 +- examples/run_multiple_choice.py | 2 +- examples/run_xnli.py | 2 +- 6 files changed, 6 insertions(+), 6 deletions(-) diff --git a/examples/hans/test_hans.py b/examples/hans/test_hans.py index 56416b28bd82..99b6f24671ea 100644 --- a/examples/hans/test_hans.py +++ b/examples/hans/test_hans.py @@ -255,7 +255,7 @@ def evaluate(args, model, tokenizer, prefix=""): eval_dataloader = DataLoader(eval_dataset, sampler=eval_sampler, batch_size=args.eval_batch_size) # multi-gpu eval - if args.n_gpu > 1: + if args.n_gpu > 1 and not isinstance(model, torch.nn.DataParallel): model = torch.nn.DataParallel(model) # Eval! diff --git a/examples/mm-imdb/run_mmimdb.py b/examples/mm-imdb/run_mmimdb.py index ab60b4095166..f74ea575b7f5 100644 --- a/examples/mm-imdb/run_mmimdb.py +++ b/examples/mm-imdb/run_mmimdb.py @@ -278,7 +278,7 @@ def evaluate(args, model, tokenizer, criterion, prefix=""): ) # multi-gpu eval - if args.n_gpu > 1: + if args.n_gpu > 1 and not isinstance(model, torch.nn.DataParallel): model = torch.nn.DataParallel(model) # Eval! diff --git a/examples/ner/run_ner.py b/examples/ner/run_ner.py index ba08e51da882..e280037ad8a3 100644 --- a/examples/ner/run_ner.py +++ b/examples/ner/run_ner.py @@ -253,7 +253,7 @@ def evaluate(args, model, tokenizer, labels, pad_token_label_id, mode, prefix="" eval_dataloader = DataLoader(eval_dataset, sampler=eval_sampler, batch_size=args.eval_batch_size) # multi-gpu evaluate - if args.n_gpu > 1: + if args.n_gpu > 1 and not isinstance(model, torch.nn.DataParallel): model = torch.nn.DataParallel(model) # Eval! diff --git a/examples/run_language_modeling.py b/examples/run_language_modeling.py index 92807dd79ca7..859bc164c335 100644 --- a/examples/run_language_modeling.py +++ b/examples/run_language_modeling.py @@ -427,7 +427,7 @@ def collate(examples: List[torch.Tensor]): ) # multi-gpu evaluate - if args.n_gpu > 1: + if args.n_gpu > 1 and not isinstance(model, torch.nn.DataParallel): model = torch.nn.DataParallel(model) # Eval! diff --git a/examples/run_multiple_choice.py b/examples/run_multiple_choice.py index 578ce0122610..341733b8a091 100644 --- a/examples/run_multiple_choice.py +++ b/examples/run_multiple_choice.py @@ -256,7 +256,7 @@ def evaluate(args, model, tokenizer, prefix="", test=False): eval_dataloader = DataLoader(eval_dataset, sampler=eval_sampler, batch_size=args.eval_batch_size) # multi-gpu evaluate - if args.n_gpu > 1: + if args.n_gpu > 1 and not isinstance(model, torch.nn.DataParallel): model = torch.nn.DataParallel(model) # Eval! diff --git a/examples/run_xnli.py b/examples/run_xnli.py index 9b64d17dfd82..7c2790f53d97 100644 --- a/examples/run_xnli.py +++ b/examples/run_xnli.py @@ -266,7 +266,7 @@ def evaluate(args, model, tokenizer, prefix=""): eval_dataloader = DataLoader(eval_dataset, sampler=eval_sampler, batch_size=args.eval_batch_size) # multi-gpu eval - if args.n_gpu > 1: + if args.n_gpu > 1 and not isinstance(model, torch.nn.DataParallel): model = torch.nn.DataParallel(model) # Eval!