Skip to content

Commit

Permalink
Apply isort
Browse files Browse the repository at this point in the history
  • Loading branch information
akihironitta committed Jan 3, 2021
1 parent c5417d6 commit 0b7c552
Show file tree
Hide file tree
Showing 44 changed files with 58 additions and 58 deletions.
4 changes: 2 additions & 2 deletions pl_bolts/callbacks/data_monitor.py
Original file line number Diff line number Diff line change
@@ -1,13 +1,13 @@
from typing import Any, Dict, List, Optional, Sequence, Union

import numpy as np
import torch
import torch.nn as nn
from pytorch_lightning import Callback, LightningModule, Trainer
from pytorch_lightning.loggers import TensorBoardLogger, WandbLogger
from pytorch_lightning.utilities import rank_zero_warn
from pytorch_lightning.utilities.apply_func import apply_to_collection
import torch
from torch import Tensor
import torch.nn as nn
from torch.utils.hooks import RemovableHandle

try:
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/callbacks/ssl_online.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
from typing import Optional

import torch
from pytorch_lightning import Callback
from pytorch_lightning.metrics.functional import accuracy
import torch
from torch.nn import functional as F


Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/callbacks/variational.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
import math

import numpy as np
from pytorch_lightning.callbacks import Callback
import torch
from pytorch_lightning.callbacks import Callback

from pl_bolts.utils.warnings import warn_missing_pkg

Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/callbacks/vision/confused_logit.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import importlib

from pytorch_lightning import Callback
import torch
from pytorch_lightning import Callback
from torch import nn

from pl_bolts.utils.warnings import warn_missing_pkg
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/callbacks/vision/image_generation.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from pytorch_lightning import Callback
import torch
from pytorch_lightning import Callback

from pl_bolts.utils.warnings import warn_missing_pkg

Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/datamodules/async_dataloader.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from queue import Queue
import re
from queue import Queue
from threading import Thread

import torch
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/datamodules/kitti_datamodule.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import os

from pytorch_lightning import LightningDataModule
import torch
from pytorch_lightning import LightningDataModule
from torch.utils.data import DataLoader
from torch.utils.data.dataset import random_split

Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/datamodules/sklearn_datamodule.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@
from typing import Any

import numpy as np
from pytorch_lightning import LightningDataModule
import torch
from pytorch_lightning import LightningDataModule
from torch.utils.data import DataLoader, Dataset

from pl_bolts.utils import _SKLEARN_AVAILABLE
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/datamodules/stl10_datamodule.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
import os
from typing import Optional

from pytorch_lightning import LightningDataModule
import torch
from pytorch_lightning import LightningDataModule
from torch.utils.data import DataLoader, random_split

from pl_bolts.datasets.concat_dataset import ConcatDataset
Expand Down
4 changes: 2 additions & 2 deletions pl_bolts/datamodules/vision_datamodule.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
from abc import abstractmethod
import os
from abc import abstractmethod
from typing import Any, List, Optional, Union

from pytorch_lightning import LightningDataModule
import torch
from pytorch_lightning import LightningDataModule
from torch.utils.data import DataLoader, Dataset, random_split


Expand Down
4 changes: 2 additions & 2 deletions pl_bolts/datamodules/vocdetection_datamodule.py
Original file line number Diff line number Diff line change
@@ -1,13 +1,13 @@
from pytorch_lightning import LightningDataModule
import torch
from pytorch_lightning import LightningDataModule
from torch.utils.data import DataLoader

from pl_bolts.utils import _TORCHVISION_AVAILABLE
from pl_bolts.utils.warnings import warn_missing_pkg

if _TORCHVISION_AVAILABLE:
from torchvision.datasets import VOCDetection
import torchvision.transforms as T
from torchvision.datasets import VOCDetection
else:
warn_missing_pkg('torchvision') # pragma: no-cover

Expand Down
4 changes: 2 additions & 2 deletions pl_bolts/datasets/base_dataset.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
from abc import ABC
import logging
import os
import urllib.request
from abc import ABC
from typing import Sequence, Tuple
from urllib.error import HTTPError
import urllib.request

import torch
from torch import Tensor
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/datasets/imagenet_dataset.py
Original file line number Diff line number Diff line change
@@ -1,11 +1,11 @@
from contextlib import contextmanager
import gzip
import hashlib
import os
import shutil
import tarfile
import tempfile
import zipfile
from contextlib import contextmanager

import numpy as np
import torch
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/autoencoders/components.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
import torch
from torch import nn
import torch.nn.functional as F
from torch import nn


class Interpolate(nn.Module):
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/gans/basic/components.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import numpy as np
import torch
from torch import nn
import torch.nn.functional as F
from torch import nn


class Generator(nn.Module):
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/mnist_module.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
from argparse import ArgumentParser

from pytorch_lightning import LightningModule, Trainer
import torch
from pytorch_lightning import LightningModule, Trainer
from torch.nn import functional as F
from torch.utils.data import DataLoader, random_split

Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/regression/logistic_regression.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
from argparse import ArgumentParser

import pytorch_lightning as pl
from pytorch_lightning.metrics.functional import accuracy
import torch
from pytorch_lightning.metrics.functional import accuracy
from torch import nn
from torch.nn import functional as F
from torch.optim import Adam
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/rl/common/agents.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,8 @@

import numpy as np
import torch
from torch import nn
import torch.nn.functional as F
from torch import nn


class Agent(ABC):
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/rl/common/gym_wrappers.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,9 +11,9 @@
from pl_bolts.utils.warnings import warn_missing_pkg

if _GYM_AVAILABLE:
import gym.spaces
from gym import make as gym_make
from gym import ObservationWrapper, Wrapper
import gym.spaces
else: # pragma: no-cover
warn_missing_pkg('gym')
Wrapper = object
Expand Down
4 changes: 2 additions & 2 deletions pl_bolts/models/rl/dqn_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,10 +8,10 @@

import numpy as np
import pytorch_lightning as pl
from pytorch_lightning import seed_everything
from pytorch_lightning.callbacks import ModelCheckpoint
import torch
import torch.optim as optim
from pytorch_lightning import seed_everything
from pytorch_lightning.callbacks import ModelCheckpoint
from torch.optim.optimizer import Optimizer
from torch.utils.data import DataLoader

Expand Down
4 changes: 2 additions & 2 deletions pl_bolts/models/rl/reinforce_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,11 +4,11 @@

import numpy as np
import pytorch_lightning as pl
import torch
import torch.optim as optim
from pytorch_lightning import seed_everything
from pytorch_lightning.callbacks import ModelCheckpoint
import torch
from torch.nn.functional import log_softmax
import torch.optim as optim
from torch.optim.optimizer import Optimizer
from torch.utils.data import DataLoader

Expand Down
4 changes: 2 additions & 2 deletions pl_bolts/models/rl/vanilla_policy_gradient_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,11 +4,11 @@

import numpy as np
import pytorch_lightning as pl
import torch
import torch.optim as optim
from pytorch_lightning import seed_everything
from pytorch_lightning.callbacks import ModelCheckpoint
import torch
from torch.nn.functional import log_softmax, softmax
import torch.optim as optim
from torch.optim.optimizer import Optimizer
from torch.utils.data import DataLoader

Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/amdim/amdim_module.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from argparse import ArgumentParser
import os
from argparse import ArgumentParser
from typing import Union

import pytorch_lightning as pl
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/amdim/networks.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@

import numpy as np
import torch
from torch import nn
import torch.nn.functional as F
from torch import nn


class AMDIMEncoder(nn.Module):
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/byol/byol_module.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,9 +3,9 @@
from typing import Any

import pytorch_lightning as pl
from pytorch_lightning import seed_everything
import torch
import torch.nn.functional as F
from pytorch_lightning import seed_everything
from torch.optim import Adam

from pl_bolts.callbacks.byol_updates import BYOLMAWeightUpdate
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/cpc/cpc_finetuner.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from argparse import ArgumentParser
import os
from argparse import ArgumentParser

import pytorch_lightning as pl

Expand Down
4 changes: 2 additions & 2 deletions pl_bolts/models/self_supervised/cpc/cpc_module.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,14 +2,14 @@
CPC V2
======
"""
from argparse import ArgumentParser
import math
from argparse import ArgumentParser
from typing import Optional

import pytorch_lightning as pl
from pytorch_lightning.utilities import rank_zero_warn
import torch
import torch.optim as optim
from pytorch_lightning.utilities import rank_zero_warn

from pl_bolts.losses.self_supervised_learning import CPCTask
from pl_bolts.models.self_supervised.cpc.networks import cpc_resnet101
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/cpc/networks.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from torch import nn
import torch.nn.functional as F
from torch import nn


class CPCResNet(nn.Module):
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/moco/moco2_module.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,8 +15,8 @@

import pytorch_lightning as pl
import torch
from torch import nn
import torch.nn.functional as F
from torch import nn

from pl_bolts.utils.warnings import warn_missing_pkg

Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/simclr/simclr_finetuner.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from argparse import ArgumentParser
import os
from argparse import ArgumentParser

import pytorch_lightning as pl

Expand Down
6 changes: 3 additions & 3 deletions pl_bolts/models/self_supervised/simclr/simclr_module.py
Original file line number Diff line number Diff line change
@@ -1,13 +1,13 @@
from argparse import ArgumentParser
import math
from argparse import ArgumentParser
from typing import Callable, Optional

import numpy as np
import pytorch_lightning as pl
from pytorch_lightning.utilities import AMPType
import torch
from torch import nn
import torch.nn.functional as F
from pytorch_lightning.utilities import AMPType
from torch import nn
from torch.optim.optimizer import Optimizer

from pl_bolts.models.self_supervised.resnets import resnet18, resnet50
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/ssl_finetuner.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
from typing import List, Optional

import pytorch_lightning as pl
from pytorch_lightning.metrics import Accuracy
import torch
import torch.nn.functional as F
from pytorch_lightning.metrics import Accuracy

from pl_bolts.models.self_supervised import SSLEvaluator

Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/self_supervised/swav/swav_finetuner.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from argparse import ArgumentParser
import os
from argparse import ArgumentParser

import pytorch_lightning as pl

Expand Down
6 changes: 3 additions & 3 deletions pl_bolts/models/self_supervised/swav/swav_module.py
Original file line number Diff line number Diff line change
@@ -1,17 +1,17 @@
"""
Adapted from official swav implementation: https://github.com/facebookresearch/swav
"""
from argparse import ArgumentParser
import math
import os
from argparse import ArgumentParser
from typing import Callable, Optional

import numpy as np
import pytorch_lightning as pl
from pytorch_lightning.utilities import AMPType
import torch
from torch import nn
import torch.distributed as dist
from pytorch_lightning.utilities import AMPType
from torch import nn
from torch.optim.optimizer import Optimizer

from pl_bolts.models.self_supervised.swav.swav_resnet import resnet18, resnet50
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/models/vision/image_gpt/igpt_module.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from argparse import ArgumentParser
import os
from argparse import ArgumentParser

import pytorch_lightning as pl
import torch
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/optimizers/lr_scheduler.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
import math
from typing import List
import warnings
from typing import List

import torch.nn as nn
from torch.optim import Adam, Optimizer
Expand Down
2 changes: 1 addition & 1 deletion pl_bolts/utils/__init__.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
from pytorch_lightning.utilities import _module_available
import torch
from pytorch_lightning.utilities import _module_available

_NATIVE_AMP_AVAILABLE: bool = _module_available("torch.cuda.amp") and hasattr(torch.cuda.amp, "autocast")

Expand Down
Loading

0 comments on commit 0b7c552

Please sign in to comment.