There’s vermin furries here?
I can neither confirm, nor deny that I am in fact D̵̡̮̻̗̖̮͔̜͈̙͖͙͍̺̀̒̍̌̑͐̓͡å̴̲͍̋̉́̀̑͊̎̐̊͡l̴̟̭̳̄̅̕͝͠͝ȩ̸͚̼̘̫̺̻̬̻̮͖̣̬̖̠̗̎̌ ̵̯͕͛́͋͌̀͝͠ͅͅG̷̛͈̩̟̟̠͓̗̘͓͍̽̒̌̔̓̈͗̐̈̿͠͠r̷̘̞̹͂̀̑̋̀͌̍͗̆͝͠͝ͅi̶̡͔͖͍̟̲̮͑̎͌̀̎b̵̡̢̹̗͔̗͍̘̣͊͊̑͒̍̑͌̽͋͌̔͝͝b̷̭̩̩̣͙̺͎̱̗͙͚̩̈́l̸̛͎̼̟̋͆͆͗̓̓̓͘͟ĺ̶̼͇͎̫̮͎̣̳͉̯̊̆̂̓̄̍̃̚e̶̢̡̛̫̣͈̺̾̅͐̾̓͒̚ͅ.̴̫̞̥̒̈̇̓́̾͗̒́̉̔͑
There’s vermin furries here?
Earlier in my career, I compiled tensorflow with CUDA/cuDNN (NVIDIA) in one container and then in another machine and container compiled with ROCm (AMD) for cancerous tissue detection in computer vision tasks. GPU acceleration in training the model was significantly more performant with NVIDIA libraries.
It’s not like you can’t train deep neural networks without NVIDIA, but their deep learning libraries combined with tensor cores in Turing-era GPUs and later make things much faster.
I will die on this hill. Their down votes mean nothing to me. I’ve seen what makes them cheer.