Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

simplify tests & cleaning #2588

Merged
merged 9 commits into from
Aug 7, 2020
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .pyrightconfig.json
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@
"pytorch_lightning/__init__.py",
"pytorch_lightning/callbacks",
"pytorch_lightning/core",
"pytorch_lightning/accelerator_backends",
"pytorch_lightning/accelerators",
"pytorch_lightning/loggers",
"pytorch_lightning/logging",
"pytorch_lightning/metrics",
Expand Down
5 changes: 0 additions & 5 deletions .run_local_tests.sh
Original file line number Diff line number Diff line change
Expand Up @@ -6,11 +6,6 @@ export SLURM_LOCALID=0

# use this to run tests
rm -rf _ckpt_*
rm -rf ./tests/save_dir*
rm -rf ./tests/mlruns_*
rm -rf ./tests/cometruns*
rm -rf ./tests/wandb*
rm -rf ./tests/tests/*
rm -rf ./lightning_logs
python -m coverage run --source pytorch_lightning -m py.test pytorch_lightning tests pl_examples -v --flake8
python -m coverage report -m
Expand Down
2 changes: 1 addition & 1 deletion docs/source/conf.py
Original file line number Diff line number Diff line change
Expand Up @@ -138,7 +138,7 @@
exclude_patterns = [
'api/pytorch_lightning.rst',
'api/pl_examples.*',
'api/pytorch_lightning.accelerator_backends.*',
'api/pytorch_lightning.accelerators.*',
'api/modules.rst',
'PULL_REQUEST_TEMPLATE.md',
]
Expand Down
7 changes: 0 additions & 7 deletions pytorch_lightning/accelerator_backends/__init__.py

This file was deleted.

7 changes: 7 additions & 0 deletions pytorch_lightning/accelerators/__init__.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
from pytorch_lightning.accelerators.gpu_backend import GPUBackend
from pytorch_lightning.accelerators.tpu_backend import TPUBackend
from pytorch_lightning.accelerators.dp_backend import DataParallelBackend
from pytorch_lightning.accelerators.ddp_spawn_backend import DDPSpawnBackend
from pytorch_lightning.accelerators.cpu_backend import CPUBackend
from pytorch_lightning.accelerators.ddp_backend import DDPBackend
from pytorch_lightning.accelerators.ddp2_backend import DDP2Backend
2 changes: 1 addition & 1 deletion pytorch_lightning/trainer/trainer.py
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@
from pytorch_lightning.utilities.debugging import InternalDebugger
from pytorch_lightning.utilities.exceptions import MisconfigurationException
from pytorch_lightning.trainer.configuration_validator import ConfigValidator
from pytorch_lightning.accelerator_backends import (
from pytorch_lightning.accelerators import (
GPUBackend, TPUBackend, CPUBackend, DDPSpawnBackend, DataParallelBackend, DDPBackend, DDP2Backend)

# warnings to ignore in trainer
Expand Down
2 changes: 1 addition & 1 deletion tests/core/test_datamodules.py
Original file line number Diff line number Diff line change
Expand Up @@ -299,7 +299,7 @@ def test_full_loop_ddp_spawn(tmpdir):

trainer = Trainer(
default_root_dir=tmpdir,
max_epochs=3,
max_epochs=5,
weights_summary=None,
distributed_backend='ddp_spawn',
gpus=[0, 1]
Expand Down
18 changes: 12 additions & 6 deletions tests/trainer/test_trainer_steps_dict_return.py
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,11 @@ def training_step_with_step_end(tmpdir):
model.training_step_end = model.training_step_end_dict
model.val_dataloader = None

trainer = Trainer(fast_dev_run=True, weights_summary=None)
trainer = Trainer(
default_root_dir=tmpdir,
fast_dev_run=True,
weights_summary=None,
)
trainer.fit(model)

# make sure correct steps were called
Expand Down Expand Up @@ -107,8 +111,7 @@ def test_full_training_loop_dict(tmpdir):
assert trainer.progress_bar_metrics['epoch_end_pbar_1'] == 234

# make sure training outputs what is expected
for batch_idx, batch in enumerate(model.train_dataloader()):
break
batch_idx, batch = 0, next(iter(model.train_dataloader()))

out = trainer.run_training_batch(batch, batch_idx)
assert out.signal == 0
Expand All @@ -131,7 +134,11 @@ def test_train_step_epoch_end(tmpdir):
model.training_epoch_end = model.training_epoch_end_dict
model.val_dataloader = None

trainer = Trainer(max_epochs=1, weights_summary=None)
trainer = Trainer(
default_root_dir=tmpdir,
max_epochs=1,
weights_summary=None,
)
trainer.fit(model)

# make sure correct steps were called
Expand All @@ -144,8 +151,7 @@ def test_train_step_epoch_end(tmpdir):
assert trainer.progress_bar_metrics['epoch_end_pbar_1'] == 234

# make sure training outputs what is expected
for batch_idx, batch in enumerate(model.train_dataloader()):
break
batch_idx, batch = 0, next(iter(model.train_dataloader()))

out = trainer.run_training_batch(batch, batch_idx)
assert out.signal == 0
Expand Down