inari@piefed.zip to Technology@lemmy.worldEnglish · 5 days agoDeepSeek ditches Nvidia for Huawei chips in V4 launchcybernews.comexternal-linkmessage-square89linkfedilinkarrow-up1442arrow-down19
arrow-up1433arrow-down1external-linkDeepSeek ditches Nvidia for Huawei chips in V4 launchcybernews.cominari@piefed.zip to Technology@lemmy.worldEnglish · 5 days agomessage-square89linkfedilink
minus-square[object Object]@lemmy.calinkfedilinkEnglisharrow-up8·5 days agoEven with a bitnet, it’s almost definitely better to train on a high precision float then refine down to bits. I would expect bitnet to require more layers for equivalent quality too.
minus-squarebrucethemoose@lemmy.worldlinkfedilinkEnglisharrow-up5arrow-down1·5 days agoI just meant for mass inference serving. Yeah, I haven’t seen much in the way of bitnet training savings yet, like regular old QAT. It does appear that Deepseek is finetuning their MoEs in a 4-bit format now, though.
Even with a bitnet, it’s almost definitely better to train on a high precision float then refine down to bits.
I would expect bitnet to require more layers for equivalent quality too.
I just meant for mass inference serving.
Yeah, I haven’t seen much in the way of bitnet training savings yet, like regular old QAT. It does appear that Deepseek is finetuning their MoEs in a 4-bit format now, though.