You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I was reading about the bitnet (1.58bit) quantization, and I was wondering if that would improve inference time for this model? That would help in the real time translation.
Let me know what do you think, and I'm happy to help with that!
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi everyone!
I was reading about the bitnet (1.58bit) quantization, and I was wondering if that would improve inference time for this model? That would help in the real time translation.
Let me know what do you think, and I'm happy to help with that!
Beta Was this translation helpful? Give feedback.
All reactions