![](https://awful.systems/pictrs/image/f4f2722f-13dc-48df-a75e-df49c67988ae.png)
![](https://awful.systems/pictrs/image/8651f454-1f76-42f4-bb27-4c64b332f07a.png)
That’s the opposite of what I’m saying. Deepseek is the one under scrutiny, yet they are the only one to publish source code and training procedures of their model.
this has absolutely fuck all to do with anything i’ve said in the slightest, but i guess you gotta toss in the talking points somewhere
e: it’s also trivially disprovable, but i don’t care if it’s actually true, i only care about headlines negative about AI
i can admit it’s possible i’m being overly cynical here and it is just sloppy journalism on Raffaele Huang/his editor/the WSJ’s part. but i still think that it’s a little suspect on the grounds that we have no idea how many times they had to restart training due to the model borking, other experiments and hidden costs, even before things like the necessary capex (which goes unmentioned in the original paper – though they note using a 2048-GPU cluster of H800’s that would put them down around $40m). i’m thinking in the mode of “the whitepaper exists to serve the company’s bottom line”
btw announcing my new V7 model that i trained for the $0.26 i found on the street just to watch the stock markets burn