• Kohaku-Blueleaf's avatar
    Add CPU fp8 support · eaa9f516
    Kohaku-Blueleaf authored
    Since norm layer need fp32, I only convert the linear operation layer(conv2d/linear)
    
    And TE have some pytorch function not support bf16 amp in CPU. I add a condition to indicate if the autocast is for unet.
    eaa9f516
devices.py 4.43 KB