NVIDIA's CUDA is a proprietary closed source parallel computing platform and application programming interface (API) that allows software to use certain types of GPUs for general purpose processing
all the major deep learning frameworks like TensorFlow, PyTorch, Caffe, Theano and MXNet added native support for CUDA GPU acceleration early on.
This created a self-reinforcing virtuous cycle — CUDA became the standard way of accessing GPU acceleration due to its popularity and support across frameworks, and frameworks aligned with it due to strong demand from users. Over time, the CUDA programming paradigm and stack became deeply embedded into all aspects of the AI ecosystem.
End date : Jan 1 2027
hmmm, how does this count if you can already export your models to ONNX today and run them on all sorts of devices? You just have to be careful what APIs you call and such, and need a portable variant if you have custom CUDA code.
But like, my Stable Diffusion and LLM trainings I did on my Mac because PyTorch just runs everywhere. And all the hardware startups are getting LLMs ported using ONNX.
@Shump Yes, that's the whole point of ONNX. Microsoft/Facebook made it as a portability layer. PyTorch team did a lot of the work.
The hardware declares to support an "ONNX version" and when you export a model it also will have an ONNX version. So not everything necessarily runs, but that's hardware/driver/library limitation. The framework is supposed to allow it.
So if e.g. Groq declares their chips support ONNX 12, then any model exported to ONNX 12 will run on a Groq card.