Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Support INT8 mixed-precision training from torchao? #578

@gau-nernst

Description

@gau-nernst

Recently I worked on INT8 mixed-precision training in torchao. The relevant PR is here pytorch/ao#748

Preliminary results show that with torchtitan, it improves speed by 20% on 8x A100 with no noticeable difference in loss curve. See the PR for more details.

Would you be open to add an experimental flag for this in torchtitan? Similar to Float8 training. This can also help to profile and improve INT8 training performance directly in torchtitan for future perf optimization.

cc @msaroufim

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions