In deep learning generally open source doesn’t include actual training or inference code. Rather it means they publish the model weights and parameters (necessary to run it locally/on your own hardware) and publish academic papers explaining how the model was trained. I’m sure Stallman disagrees but from the standpoint of deep learning research DeepSeek definitely qualifies as an “open source model”
Obviously you need lots of GPUs to run large deep learning models. I don’t see how that’s a fault of the developers and researchers, it’s just a fact of this technology.