Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's not quantising existing models, they're training new ones.


I understand this part but it seemed that the 16->8->4 etc is similar to compression of the "net" and seemed to lower quality below 8.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: