

- open weights, as the training dataset is not open/available afaik. But yum :D
One thing to be kept in mind, though:
verified this myself with the 1.1b model
Thanks for clarification!
So… as far as I understand from this thread, it’s basically a finished model (llama or qwen) which is then fine tuned using an unknown dataset? That’d explain the claimed 6M training cost, hiding the fact that the heavy lifting has been made by others (US of A’s Meta in this case). Nothing revolutionary to see here, I guess. Small improvements are nice to have, though. I wonder how their smallest models perform, are they any better than llama3.2:8b?
why are you so heavily and openly advertising Deepseek?
Mind elaborating? I’m interested in anything keeping me away from excel
Meanwhile MS EXCEL: This file with random macros from a shady website could gain admin rights, install 3500 viruses, lock you out, join a botnet, put a million dollar ransom on your PC all within the first minute after opening without you even noticing. Please click ok if you are fine with that.
Really wonder how they plan to increase their revenue on the AI training data, especially now that a significant amount of their data is “poisoned” by the models they try to train
We are not the same
and both tired as fuck.
Seriously, why am I as tired after 14 hours of sleep as if I slept for 6 hours?
About time they rebrand as ClosedAI.
Is it a war or a cartel, though?