Skip to content

Commit 7d65dd9

Browse files
committed
Update GH org references
1 parent 241bffd commit 7d65dd9

File tree

125 files changed

+404
-404
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

125 files changed

+404
-404
lines changed

.github/ISSUE_TEMPLATE/deepspeed_chat_bug_report.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ If applicable, add screenshots to help explain your problem.
3232
**System info (please complete the following information):**
3333
- OS: [e.g. Ubuntu 18.04]
3434
- GPU count and types [e.g. two machines with x8 A100s each]
35-
- (if applicable) what [DeepSpeed-MII](https://github.com/microsoft/deepspeed-mii) version are you using
35+
- (if applicable) what [DeepSpeed-MII](https://github.com/deepspeedai/deepspeed-mii) version are you using
3636
- (if applicable) Hugging Face Transformers/Accelerate/etc. versions
3737
- Python version
3838
- Any other relevant info about your setup

.github/workflows/nv-a6000.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -66,7 +66,7 @@ jobs:
6666
BRANCH="${{ github.event.inputs.mii_branch }}"
6767
fi
6868
echo "Cloning DeepSpeed-MII branch: $BRANCH"
69-
git clone -b $BRANCH --depth=1 https://github.com/microsoft/DeepSpeed-MII.git
69+
git clone -b $BRANCH --depth=1 https://github.com/deepspeedai/DeepSpeed-MII.git
7070
cd DeepSpeed-MII
7171
pip install .[dev]
7272
cd tests

.github/workflows/nv-ds-chat.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ jobs:
5454
BRANCH="${{ github.event.inputs.dse_branch }}"
5555
fi
5656
echo "DeepSpeedExamples Branch: $BRANCH"
57-
git clone -b $BRANCH https://github.com/microsoft/DeepSpeedExamples.git
57+
git clone -b $BRANCH https://github.com/deepspeedai/DeepSpeedExamples.git
5858
cd DeepSpeedExamples/applications/DeepSpeed-Chat
5959
pip install -r requirements.txt
6060
pip install -e .

.github/workflows/nv-mii.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -66,7 +66,7 @@ jobs:
6666
BRANCH="${{ github.event.inputs.mii_branch }}"
6767
fi
6868
echo "Cloning DeepSpeed-MII branch: $BRANCH"
69-
git clone -b $BRANCH --depth=1 https://github.com/microsoft/DeepSpeed-MII.git
69+
git clone -b $BRANCH --depth=1 https://github.com/deepspeedai/DeepSpeed-MII.git
7070
cd DeepSpeed-MII
7171
pip install .[dev]
7272
unset TORCH_CUDA_ARCH_LIST # only jit compile for current arch

CONTRIBUTING.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ and then repeat the previous `git commit` command.
2323
## Testing
2424
DeepSpeed tracks two types of tests: unit tests and more costly model convergence tests.
2525
The model convergence tests train
26-
[DeepSpeedExamples](https://github.com/microsoft/DeepSpeedExamples/) and measure
26+
[DeepSpeedExamples](https://github.com/deepspeedai/DeepSpeedExamples/) and measure
2727
end-to-end convergence and related metrics. Unit tests are found in `tests/unit/` and
2828
the model convergence tests are found in `tests/model/`.
2929

@@ -40,7 +40,7 @@ tests. Note that [pytest-forked](https://github.com/pytest-dev/pytest-forked) an
4040

4141
### Model Tests
4242
To execute model tests, first [install DeepSpeed](#installation). The
43-
[DeepSpeedExamples](https://github.com/microsoft/DeepSpeedExamples/) repository is cloned
43+
[DeepSpeedExamples](https://github.com/deepspeedai/DeepSpeedExamples/) repository is cloned
4444
as part of this process. Next, execute the model test driver:
4545
```bash
4646
cd tests/model/
@@ -85,8 +85,8 @@ Based on the issue we shall discuss the merit of the new feature and decide whet
8585
### Step 2: implementation and verification
8686
Contributor will go ahead and implement the feature, and the DeepSpeed team will provide guidance/helps as needed. The required deliverables include:
8787

88-
* A PR to [microsoft/DeepSpeed](https://github.com/microsoft/DeepSpeed) including (1) the feature implementation (2) unit tests (3) documentation (4) tutorial
89-
* A PR to [microsoft/DeepSpeedExamples](https://github.com/microsoft/DeepSpeedExamples) or [microsoft/Megatron-DeepSpeed](https://github.com/microsoft/Megatron-DeepSpeed) including the examples of how to use the feature (this is related to the planned testing experiments in proposal)
88+
* A PR to [deepspeedai/DeepSpeed](https://github.com/deepspeedai/DeepSpeed) including (1) the feature implementation (2) unit tests (3) documentation (4) tutorial
89+
* A PR to [deepspeedai/DeepSpeedExamples](https://github.com/deepspeedai/DeepSpeedExamples) or [deepspeedai/Megatron-DeepSpeed](https://github.com/deepspeedai/Megatron-DeepSpeed) including the examples of how to use the feature (this is related to the planned testing experiments in proposal)
9090
* In the implementation (code, documentation, tutorial), we require the feature author to record their GitHub username as a contact method for future questions/maintenance.
9191

9292
After receiving the PRs, we will review them and merge them after necessary tests/fixes.

README.md

Lines changed: 34 additions & 34 deletions
Large diffs are not rendered by default.

accelerator/real_accelerator.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -169,7 +169,7 @@ def get_accelerator():
169169
# ensuring that we are free from CUDA initialization errors.
170170
# While "torch.cuda.device_count() > 0" check ensures that #ignore-cuda
171171
# we won't try to do any CUDA calls when no device is available
172-
# For reference: https://github.com/microsoft/DeepSpeed/pull/6810
172+
# For reference: https://github.com/deepspeedai/DeepSpeed/pull/6810
173173
if torch.cuda.device_count() > 0 and torch.cuda.is_available(): #ignore-cuda
174174
accelerator_name = "cuda"
175175
except (RuntimeError, ImportError) as e:

benchmarks/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,5 +2,5 @@
22

33
If you are looking for DeepSpeed benchmarks, please see the following resources:
44

5-
1. [Communication Benchmarking Suite](https://github.com/microsoft/DeepSpeedExamples/tree/master/benchmarks/communication)
6-
2. [Inference Benchmarks](https://github.com/microsoft/DeepSpeedExamples/tree/master/benchmarks/inference)
5+
1. [Communication Benchmarking Suite](https://github.com/deepspeedai/DeepSpeedExamples/tree/master/benchmarks/communication)
6+
2. [Inference Benchmarks](https://github.com/deepspeedai/DeepSpeedExamples/tree/master/benchmarks/inference)

blogs/deepspeed-chat/README.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ DeepSpeed-RLHF system is capable of unparalleled efficiency at scale, making com
6565
*Table 2. Multi-Node 64x A100-80GB: Training Time and Corresponding Approximate Cost on Azure.*
6666
</div>
6767

68-
> ***Very Important Details***: The numbers in both tables above are for Step 3 of the training and are based on actual measured training throughput on DeepSpeed-RLHF curated dataset and training recipe which trains for one epoch on a total of 135M tokens. We have in total 67.5M query tokens (131.9k queries with sequence length 256) and 67.5M generated tokens (131.9k answers with sequence length 256), and a maximum global batch size per step of 0.5M tokens (1024 query-answer pairs). We urge readers to pay attention to these specifications before making any cost and e2e time comparisons with DeepSpeed-RLHF. See our [benchmark settings](https://github.com/microsoft/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat/training/step3_rlhf_finetuning/BenckmarkSetting.md) page for more details.
68+
> ***Very Important Details***: The numbers in both tables above are for Step 3 of the training and are based on actual measured training throughput on DeepSpeed-RLHF curated dataset and training recipe which trains for one epoch on a total of 135M tokens. We have in total 67.5M query tokens (131.9k queries with sequence length 256) and 67.5M generated tokens (131.9k answers with sequence length 256), and a maximum global batch size per step of 0.5M tokens (1024 query-answer pairs). We urge readers to pay attention to these specifications before making any cost and e2e time comparisons with DeepSpeed-RLHF. See our [benchmark settings](https://github.com/deepspeedai/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat/training/step3_rlhf_finetuning/BenckmarkSetting.md) page for more details.
6969
7070

7171
***Democratizing RLHF Training***: With just a single GPU, DeepSpeed-HE supports training models with over 13 billion parameters, enabling data scientists without access to multi-GPU systems to create not just toy RLHF models but large and powerful ones that can be used in real-world scenarios.
@@ -95,7 +95,7 @@ We use an example of pretrained OPT-13B as the actor model and OPT-350M as the r
9595
```python
9696
pip install deepspeed>=0.9.0
9797

98-
git clone https://github.com/microsoft/DeepSpeedExamples.git
98+
git clone https://github.com/deepspeedai/DeepSpeedExamples.git
9999
cd DeepSpeedExamples/applications/DeepSpeed-Chat/
100100
pip install -r requirements.txt
101101

@@ -285,7 +285,7 @@ This improvement in efficiency stems from DeepSpeed-HE’s ability to accelerate
285285

286286
## Effective Throughput and Scalability Analysis
287287

288-
***(I) Effective Throughput Analysis.*** The effective throughput of DeepSpeed-HE during Stage 3 of the RLHF training depends on the throughput that it achieves during the generation and RL training phases. In our RLHF pipeline, the generation phase comprises approximately 20% of the total computation while the RL training phase comprises of remaining 80% (see [benchmark settings](https://github.com/microsoft/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat/training/step3_rlhf_finetuning/BenckmarkSetting.md) page for details). However, despite having a small proportion, the former can take a large portion of the e2e time as it requires running the actor model once for each of the 256 generated tokens with an initial prompt of 256 tokens, making it memory bandwidth bound and difficult to achieve high throughput. In contrast, the RL training phase is compute bound running the reference actor model with just a couple of forward and backward passes with full 512 tokens from both prompt and generation per sample and can achieve good throughput.
288+
***(I) Effective Throughput Analysis.*** The effective throughput of DeepSpeed-HE during Stage 3 of the RLHF training depends on the throughput that it achieves during the generation and RL training phases. In our RLHF pipeline, the generation phase comprises approximately 20% of the total computation while the RL training phase comprises of remaining 80% (see [benchmark settings](https://github.com/deepspeedai/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat/training/step3_rlhf_finetuning/BenckmarkSetting.md) page for details). However, despite having a small proportion, the former can take a large portion of the e2e time as it requires running the actor model once for each of the 256 generated tokens with an initial prompt of 256 tokens, making it memory bandwidth bound and difficult to achieve high throughput. In contrast, the RL training phase is compute bound running the reference actor model with just a couple of forward and backward passes with full 512 tokens from both prompt and generation per sample and can achieve good throughput.
289289

290290
<div align="center">
291291

@@ -320,13 +320,13 @@ As a result, for a given max global batch size, DeepSpeed-HE achieves the best t
320320

321321
We are very excited to share that DeepSpeed-Chat is now open-sourced and available to the AI community.
322322

323-
* To get started, please visit our github page for DeepSpeed-Chat: [GitHub Landing Page](https://github.com/microsoft/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat)
323+
* To get started, please visit our github page for DeepSpeed-Chat: [GitHub Landing Page](https://github.com/deepspeedai/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat)
324324

325-
* We will continue to improve DeepSpeed-Chat with your feedback and support. Our [roadmap](https://github.com/microsoft/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat/README.md#-deepspeed-chats-roadmap-) shows currently supported features as well as ones that are planned for the future.
325+
* We will continue to improve DeepSpeed-Chat with your feedback and support. Our [roadmap](https://github.com/deepspeedai/DeepSpeedExamples/tree/master/applications/DeepSpeed-Chat/README.md#-deepspeed-chats-roadmap-) shows currently supported features as well as ones that are planned for the future.
326326

327327
DeepSpeed-Chat is part of the bigger DeepSpeed ecosystem comprising a multitude of Deep Learning systems and modeling technologies. To learn more,
328328

329329
* Please visit our [website](https://www.deepspeed.ai/) for detailed blog posts, tutorials, and helpful documentation.
330330
* You can also follow us on our [English Twitter](https://twitter.com/MSFTDeepSpeed), [Japanese Twitter](https://twitter.com/MSFTDeepSpeedJP), and [Chinese Zhihu](https://www.zhihu.com/people/deepspeed) for latest news on DeepSpeed.
331331

332-
DeepSpeed welcomes your contributions! We encourage you to report issues, contribute PRs, and join discussions on the [DeepSpeed GitHub](https://github.com/microsoft/DeepSpeed/) page. Please see our [contributing guide](https://github.com/microsoft/DeepSpeed/blob/master/CONTRIBUTING.md) for more details. We are open to collaborations with universities, research labs, companies, such as those working together on deep learning research, applying DeepSpeed to empower real-world AI models and applications, and so on. For such requests (and other requests unsuitable for GitHub), please directly email to deepspeed-info@microsoft.com.
332+
DeepSpeed welcomes your contributions! We encourage you to report issues, contribute PRs, and join discussions on the [DeepSpeed GitHub](https://github.com/deepspeedai/DeepSpeed/) page. Please see our [contributing guide](https://github.com/deepspeedai/DeepSpeed/blob/master/CONTRIBUTING.md) for more details. We are open to collaborations with universities, research labs, companies, such as those working together on deep learning research, applying DeepSpeed to empower real-world AI models and applications, and so on. For such requests (and other requests unsuitable for GitHub), please directly email to deepspeed-info@microsoft.com.

0 commit comments

Comments
 (0)