Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support INT8 mixed-precision training from torchao? #578

Open
gau-nernst opened this issue Sep 14, 2024 · 0 comments
Open

Support INT8 mixed-precision training from torchao? #578

gau-nernst opened this issue Sep 14, 2024 · 0 comments

Comments

@gau-nernst
Copy link

Recently I worked on INT8 mixed-precision training in torchao. The relevant PR is here pytorch/ao#748

Preliminary results show that with torchtitan, it improves speed by 20% on 8x A100 with no noticeable difference in loss curve. See the PR for more details.

Would you be open to add an experimental flag for this in torchtitan? Similar to Float8 training. This can also help to profile and improve INT8 training performance directly in torchtitan for future perf optimization.

cc @msaroufim

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant