Abstract: To date, the most promising methods for 8-bit DNN training use two different floating-point formats: a 5-bit ex-ponent for greater range on gradients in the backwards pass, and a 4-bit ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results