Ah yeah those were the good old days when vendors were free to do that, before AMD/Nvidia restricted them. It wasn’t even that long ago, I remember some AMD 7970s being double VRAM.
And, again, I’d like to point out how insane this restriction is for AMD given their market struggles…
And a small cluster like Alibaba used to train Qwen 2.5 is basically a drop in the bucket.
The hoard of GPUs Meta, Microsoft/OpenAI, and especially X have are apparently being used extremely inefficiently, or perhaps mostly not used to train AI at all, but do regular ad/engagement optimization stuff.