Apparently I am an idiot and read the wrong paper. The previous paper mentioned that “comparable with the 8-bit models”
They said their’s is “comparable with the 8-bit models”. Its all tradeoffs. It isn’t clear to me where you allocate your compute/memory budget. I’ve noticed that full 7b 16 bit models often produce better results for me than some much larger quantied models. It will be interesting to find the sweet spot.
So are more bits less important than more paramters? Would a higher paramter or higher bit count matter more if the models ended up the same size?
I love this channel so much
Gamemaker is still focused on 2D. I also think its easier for people without programming experience and the design guides you more but is less flexible.
That might be hard unless they start requiring logins.
Firmware is the Hardware of Software
If you don’t have a good GPU then just use gpt4all
How does the Nemo 12B compare to the Llama 3.1 8B?