Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

autotp training(fix dco) #7004

Merged
merged 1 commit into from
Feb 5, 2025

Conversation

inkcherry
Copy link
Contributor

@inkcherry inkcherry commented Feb 5, 2025

Same as this PR. affeb88
I noticed the CI updated the DCO check recently. Using the suggested rebase method for sign-off would reintroduce many conflicts, so I opted for a squash merge with sign-off instead. thanks: )

Signed-off-by: inkcherry <mingzhi.liu@intel.com>
@tjruwase tjruwase added this pull request to the merge queue Feb 5, 2025
Merged via the queue into deepspeedai:master with commit f04649d Feb 5, 2025
12 checks passed
@delock
Copy link
Collaborator

delock commented Feb 6, 2025

Kudos @inkcherry for contributing AutoTP training! It's a nice feature make tensor parallel training/finetuning more available to HF model users.

I think a tutorial page would help user discover and learn how to use this feature in DeepSpeed. Is it possible to write a tutorial and add it under https://github.com/deepspeedai/DeepSpeed/tree/master/docs/_tutorials introducing steps how to use this feature? I remember you have an example training alpaca with DeepSpeed AutoTP.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants