Skip to content

Commit e260f39

Browse files
committed
rename org Lightning AI
1 parent ae3f731 commit e260f39

File tree

15 files changed

+32
-32
lines changed

15 files changed

+32
-32
lines changed

pl_examples/basic_examples/mnist_datamodule.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ class _MNIST(Dataset):
3636
"""Carbon copy of ``tests.helpers.datasets.MNIST``.
3737
3838
We cannot import the tests as they are not distributed with the package.
39-
See https://github.com/PyTorchLightning/pytorch-lightning/pull/7614#discussion_r671183652 for more context.
39+
See https://github.com/Lightning-AI/lightning/pull/7614#discussion_r671183652 for more context.
4040
"""
4141

4242
RESOURCES = (

pl_examples/domain_templates/reinforce_learn_ppo.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -125,7 +125,7 @@ def get_log_prob(self, pi: Normal, actions: torch.Tensor):
125125

126126

127127
class ExperienceSourceDataset(IterableDataset):
128-
"""Implementation from PyTorch Lightning Bolts: https://github.com/PyTorchLightning/lightning-
128+
"""Implementation from PyTorch Lightning Bolts: https://github.com/Lightning-AI/lightning-
129129
bolts/blob/master/pl_bolts/datamodules/experience_source.py.
130130
131131
Basic experience source dataset. Takes a generate_batch function that returns an iterator. The logic for the

pytorch_lightning/loggers/neptune.py

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@
5454

5555
log = logging.getLogger(__name__)
5656

57-
_INTEGRATION_VERSION_KEY = "source_code/integrations/pytorch-lightning"
57+
_INTEGRATION_VERSION_KEY = "source_code/integrations/lightning"
5858

5959
# kwargs used in previous NeptuneLogger version, now deprecated
6060
_LEGACY_NEPTUNE_INIT_KWARGS = [
@@ -113,7 +113,7 @@ class NeptuneLogger(LightningLoggerBase):
113113
114114
neptune_logger = NeptuneLogger(
115115
api_key="ANONYMOUS", # replace with your own
116-
project="common/pytorch-lightning-integration", # format "<WORKSPACE/PROJECT>"
116+
project="common/lightning-integration", # format "<WORKSPACE/PROJECT>"
117117
tags=["training", "resnet"], # optional
118118
)
119119
trainer = Trainer(max_epochs=10, logger=neptune_logger)
@@ -157,7 +157,7 @@ def any_lightning_module_function_or_hook(self):
157157
158158
.. code-block:: python
159159
160-
neptune_logger = NeptuneLogger(project="common/pytorch-lightning-integration")
160+
neptune_logger = NeptuneLogger(project="common/lightning-integration")
161161
162162
trainer = pl.Trainer(logger=neptune_logger)
163163
model = ...
@@ -182,7 +182,7 @@ def any_lightning_module_function_or_hook(self):
182182
183183
.. code-block:: python
184184
185-
neptune_logger = NeptuneLogger(project="common/pytorch-lightning-integration", log_model_checkpoints=False)
185+
neptune_logger = NeptuneLogger(project="common/lightning-integration", log_model_checkpoints=False)
186186
187187
**Pass additional parameters to the Neptune run**
188188
@@ -194,7 +194,7 @@ def any_lightning_module_function_or_hook(self):
194194
from pytorch_lightning.loggers import NeptuneLogger
195195
196196
neptune_logger = NeptuneLogger(
197-
project="common/pytorch-lightning-integration",
197+
project="common/lightning-integration",
198198
name="lightning-run",
199199
description="mlp quick run with pytorch-lightning",
200200
tags=["mlp", "quick-run"],
@@ -216,10 +216,10 @@ def any_lightning_module_function_or_hook(self):
216216
See Also:
217217
- Read about
218218
`what object you can log to Neptune <https://docs.neptune.ai/you-should-know/what-can-you-log-and-display>`_.
219-
- Check `example run <https://app.neptune.ai/o/common/org/pytorch-lightning-integration/e/PTL-1/all>`_
219+
- Check `example run <https://app.neptune.ai/o/common/org/lightning-integration/e/PTL-1/all>`_
220220
with multiple types of metadata logged.
221221
- For more detailed info check
222-
`user guide <https://docs.neptune.ai/integrations-and-supported-tools/model-training/pytorch-lightning>`_.
222+
`user guide <https://docs.neptune.ai/integrations-and-supported-tools/model-training/lightning>`_.
223223
224224
Args:
225225
api_key: Optional.
@@ -350,7 +350,7 @@ def _verify_input_arguments(
350350
" - https://docs-legacy.neptune.ai/integrations/pytorch_lightning.html\n"
351351
"The NeptuneLogger was re-written to use the neptune.new Python API\n"
352352
" - https://neptune.ai/blog/neptune-new\n"
353-
" - https://docs.neptune.ai/integrations-and-supported-tools/model-training/pytorch-lightning\n"
353+
" - https://docs.neptune.ai/integrations-and-supported-tools/model-training/lightning\n"
354354
"You should use arguments accepted by either NeptuneLogger.init() or neptune.init()"
355355
)
356356

@@ -377,7 +377,7 @@ def _verify_input_arguments(
377377
" - https://docs-legacy.neptune.ai/integrations/pytorch_lightning.html\n"
378378
"The NeptuneLogger was re-written to use the neptune.new Python API\n"
379379
" - https://neptune.ai/blog/neptune-new\n"
380-
" - https://docs.neptune.ai/integrations-and-supported-tools/model-training/pytorch-lightning\n"
380+
" - https://docs.neptune.ai/integrations-and-supported-tools/model-training/lightning\n"
381381
)
382382

383383
# check if user passed redundant neptune.init arguments when passed run
@@ -477,7 +477,7 @@ def log_hyperparams(self, params: Union[Dict[str, Any], Namespace]) -> None: #
477477
478478
neptune_logger = NeptuneLogger(
479479
api_key="ANONYMOUS",
480-
project="common/pytorch-lightning-integration"
480+
project="common/lightning-integration"
481481
)
482482
483483
neptune_logger.log_hyperparams(PARAMS)
@@ -627,7 +627,7 @@ def _signal_deprecated_api_usage(f_name, sample_code, raise_exception=False):
627627
f" - https://docs-legacy.neptune.ai/integrations/pytorch_lightning.html\n"
628628
f"The NeptuneLogger was re-written to use the neptune.new Python API\n"
629629
f" - https://neptune.ai/blog/neptune-new\n"
630-
f" - https://docs.neptune.ai/integrations-and-supported-tools/model-training/pytorch-lightning\n"
630+
f" - https://docs.neptune.ai/integrations-and-supported-tools/model-training/lightning\n"
631631
f"Instead of `logger.{f_name}` you can use:\n"
632632
f"\t{sample_code}"
633633
)

pytorch_lightning/loops/epoch/evaluation_epoch_loop.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -195,7 +195,7 @@ def _reload_dataloader_state_dict(self, data_fetcher: AbstractDataFetcher) -> No
195195
if isinstance(dataloader, CombinedLoader):
196196
raise MisconfigurationException(
197197
"Reloading support hasn't been implemented for `CombinedLoader`. You can request it by opening an issue"
198-
" in `https://github.com/PyTorchLightning/pytorch-lightning/issues`."
198+
" in `https://github.com/Lightning-AI/lightning/issues`."
199199
)
200200
assert isinstance(dataloader, DataLoader)
201201
_reload_dataloader_state_dict(dataloader, self._dataloader_state_dict)

pytorch_lightning/plugins/io/torch_plugin.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ def save_checkpoint(self, checkpoint: Dict[str, Any], path: _PATH, storage_optio
5454
atomic_save(checkpoint, path)
5555
except AttributeError as err:
5656
# todo (sean): is this try catch necessary still?
57-
# https://github.com/PyTorchLightning/pytorch-lightning/pull/431
57+
# https://github.com/Lightning-AI/lightning/pull/431
5858
key = pl.LightningModule.CHECKPOINT_HYPER_PARAMS_KEY
5959
checkpoint.pop(key, None)
6060
rank_zero_warn(f"Warning, `{key}` dropped from checkpoint. An attribute is not picklable: {err}")

pytorch_lightning/plugins/precision/ipu.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ def optimizer_step(
7272
# we lack coverage here and IPUs are (currently) limited - something to explore if there's demand
7373
raise MisconfigurationException(
7474
"Skipping backward by returning `None` from your `training_step` is not implemented for IPUs."
75-
" Please, open an issue in `https://github.com/PyTorchLightning/pytorch-lightning/issues`"
75+
" Please, open an issue in `https://github.com/Lightning-AI/lightning/issues`"
7676
" requesting this feature."
7777
)
7878
return closure_result

pytorch_lightning/plugins/precision/tpu.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@ def optimizer_step(
4646
# we lack coverage here so disable this - something to explore if there's demand
4747
raise MisconfigurationException(
4848
"Skipping backward by returning `None` from your `training_step` is not implemented for TPUs."
49-
" Please, open an issue in `https://github.com/PyTorchLightning/pytorch-lightning/issues`"
49+
" Please, open an issue in `https://github.com/Lightning-AI/lightning/issues`"
5050
" requesting this feature."
5151
)
5252
return closure_result

pytorch_lightning/setup_tools.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ def _load_readme_description(path_dir: str, homepage: str, version: str) -> str:
5959
# drop images from readme
6060
text = text.replace("![PT to PL](docs/source/_static/images/general/pl_quick_start_full_compressed.gif)", "")
6161

62-
# https://github.com/PyTorchLightning/pytorch-lightning/raw/master/docs/source/_static/images/lightning_module/pt_to_pl.png
62+
# https://github.com/Lightning-AI/lightning/raw/master/docs/source/_static/images/lightning_module/pt_to_pl.png
6363
github_source_url = os.path.join(homepage, "raw", version)
6464
# replace relative repository path to absolute link to the release
6565
# do not replace all "docs" as in the readme we reger some other sources with particular path to docs
@@ -81,7 +81,7 @@ def _load_readme_description(path_dir: str, homepage: str, version: str) -> str:
8181
# todo: wrap content as commented description
8282
text = re.sub(rf"{skip_begin}.+?{skip_end}", "<!-- -->", text, flags=re.IGNORECASE + re.DOTALL)
8383

84-
# # https://github.com/Borda/pytorch-lightning/releases/download/1.1.0a6/codecov_badge.png
84+
# # https://github.com/Borda/lightning/releases/download/1.1.0a6/codecov_badge.png
8585
# github_release_url = os.path.join(homepage, "releases", "download", version)
8686
# # download badge and replace url with local file
8787
# text = _parse_for_badge(text, github_release_url)

pytorch_lightning/strategies/launchers/spawn.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ def __init__(self, strategy: Strategy) -> None:
5353
def is_interactive_compatible(self) -> bool:
5454
# The start method 'spawn' is currently the only one that works with DDP and CUDA support
5555
# The start method 'fork' is the only one supported in Jupyter environments but not compatible with CUDA
56-
# For more context, see https://github.com/PyTorchLightning/pytorch-lightning/issues/7550
56+
# For more context, see https://github.com/Lightning-AI/lightning/issues/7550
5757
return self._start_method == "fork" and self._strategy.root_device.type != "cuda"
5858

5959
def launch(self, function: Callable, *args: Any, trainer: Optional["pl.Trainer"] = None, **kwargs: Any) -> Any:

pytorch_lightning/trainer/connectors/accelerator_connector.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -221,7 +221,7 @@ def _init_deterministic(self, deterministic: Optional[bool]) -> None:
221221
torch.use_deterministic_algorithms(self.deterministic)
222222
if self.deterministic:
223223
# fixing non-deterministic part of horovod
224-
# https://github.com/PyTorchLightning/pytorch-lightning/pull/1572/files#r420279383
224+
# https://github.com/Lightning-AI/lightning/pull/1572/files#r420279383
225225
os.environ["HOROVOD_FUSION_THRESHOLD"] = "0"
226226

227227
# https://docs.nvidia.com/cuda/cublas/index.html#cublasApi_reproducibility
@@ -605,7 +605,7 @@ def _check_strategy_and_fallback(self) -> None:
605605
if _TPU_AVAILABLE:
606606
raise MisconfigurationException(
607607
"`accelerator='ddp_cpu'` is not supported on TPU machines. "
608-
"Learn more: https://github.com/PyTorchLightning/pytorch-lightning/issues/7810"
608+
"Learn more: https://github.com/Lightning-AI/lightning/issues/7810"
609609
)
610610
if self._devices_flag == 1 and self._num_nodes_flag > 1:
611611
strategy_flag = DDPStrategy.strategy_name
@@ -725,7 +725,7 @@ def _validate_precision_choice(self) -> None:
725725
if self._precision_flag == 64:
726726
raise MisconfigurationException(
727727
"`Trainer(accelerator='tpu', precision=64)` is not implemented."
728-
" Please, open an issue in `https://github.com/PyTorchLightning/pytorch-lightning/issues`"
728+
" Please, open an issue in `https://github.com/Lightning-AI/lightning/issues`"
729729
" requesting this feature."
730730
)
731731
if self._precision_plugin_flag and not isinstance(

0 commit comments

Comments
 (0)