You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the paper, it says using w4a4 quantization can theoretically produce 8x inference speedup. Could you please confirm this for SD or what sort of speedup (inference latency) you observed? Thanks
The text was updated successfully, but these errors were encountered:
In the paper, it says using w4a4 quantization can theoretically produce 8x inference speedup. Could you please confirm this for SD or what sort of speedup (inference latency) you observed? Thanks
The text was updated successfully, but these errors were encountered: