Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Allow bfloat16 computations on compatible CPUs with Intel Extension for PyTorch #3649

Open
wants to merge 1 commit into
base: master
Choose a base branch
from

Commits on Aug 6, 2024

  1. Allow bf16 computations on CPUs with BF16 support

    Modern CPUs have native AVX512 BF16 instructions, which significantly improves
    matmul and conv2d operations.
    
    With Bfloat16 instructions UNET steps are 40-50% faster on both AMD and Intel CPUs.
    There are minor visible changes with bf16, but no avalanche effects, so this feature
    is enabled by default with new `--use-cpu-bf16=auto` option.
    It can be disabled with `--use-cpu-bf16=no`.
    
    Signed-off-by: Sv. Lockal <[email protected]>
    AngryLoki committed Aug 6, 2024
    Configuration menu
    Copy the full SHA
    88f3f92 View commit details
    Browse the repository at this point in the history