Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update references to new GitHub org (deepspeedai) #461

Closed
wants to merge 21 commits into from

Conversation

loadams
Copy link

@loadams loadams commented Feb 7, 2025

No description provided.

lekurile and others added 21 commits February 7, 2025 09:34
This PR adds a Llama universal checkpointing example to examples_deepspeed/universal_checkpointing.

It also includes changes to the README, some minor changes, and an update to the TensorBoard analysis script.

Signed-off-by: Logan Adams <[email protected]>
…sing flash_attn_cuda in sequence parallel (#406)

Co-authored-by: Jinghan Yao <[email protected]>
Signed-off-by: Logan Adams <[email protected]>
…on for supporting batch size larger than 1 (#433)

Co-authored-by: Jinghan Yao <[email protected]>
Signed-off-by: Logan Adams <[email protected]>
* add support converting checkpoint from hf to mds

* Fix PP issue

* update

Signed-off-by: Logan Adams <[email protected]>
* fix TFLOPs calculation

when GQA used, we observe right TFLOPs after this fix.
when GQA is not used, huge difference in TFLOPs is solved with 
selective recompute .
some other minor difference will also be observed as logits macs also added.

* add copyrights

Signed-off-by: Logan Adams <[email protected]>
Co-authored-by: Logan Adams <[email protected]>
Signed-off-by: Logan Adams <[email protected]>
Signed-off-by: Logan Adams <[email protected]>
* pass batch_dim_idx to deepspeed sequence parallel distributed attention for supporting batch size larger than 1

* add FPDT support; add Ulysses rotary position embedding support

* add FPDT support; add Ulysses rotary position embedding support

* add FPDT support; add Ulysses rotary position embedding support

* add FPDT support; add Ulysses rotary position embedding support

* remove unnecessary files

* set the warmup length to be FPDT chunk size if enabled

---------

Co-authored-by: Jinghan Yao <[email protected]>
Co-authored-by: Jinghan Yao <[email protected]>
Signed-off-by: Logan Adams <[email protected]>
* [tools]GQA convert support

* fix readme

Signed-off-by: Logan Adams <[email protected]>
Previously, `deepspeed_to_megatron.py` would raise an import error
due to the relative import.

This commit fixes this issue by changing from the relative import
to the absolute import like in `deepspeed_to_transformers.py`.

Signed-off-by: Logan Adams <[email protected]>
Signed-off-by: Logan Adams <[email protected]>
@loadams loadams force-pushed the loadams/update-gh-org branch from 73a7b40 to 6f50508 Compare February 7, 2025 17:36
@loadams loadams closed this Feb 7, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

10 participants