I thought someone use threadripper. No one uses it?
With pytorch and cuml you can push all your ML load to GPUs. They are faster and cheaper then a threadripper.
Thank you very much. I misunderstood that for GBDT many core CPU is faster than GPU. I will tri it.
I have NVLINK on both my machines. I havenāt used it for Numerai data, although I have for used it for some large computer vision models in the past. I donāt know of any automatic way to use multiple GPUs as one, for this use-case (model parallelism). When people say multi-GPU training, they usually mean training large batches on multiple GPUs (data parallelism), which is the easy case and trivially automated by all frameworks.
Itās fairly straightforward to place some layers (i.e weights for those layers) on different GPUs. Iāve done this with tensorflow and JAX. Still a bit slower than using a single GPU, but noticeably faster than when doing the same without NVLINK, because device to device copies are much faster with it.
My CPU is a first Gen threadripper. 12 cores 24 threads, and fairly slow compared to latest Gen CPUs (Iāve overlooked it though so all cores are running at 4ghz all the time.). I use it mostly for pre and post processing. Been contemplating if I should upgrade to a 2990wx second Gen (32 cores 64 threads), which is the biggest one the motherboard supports. I donāt really need it though.
Thanks for a very thorough answer!
3090 x 2⦠Do the house lights dim just a little when you fire that up? Nice setup!
An Asus Eee PC 901, shipped with an Intel N270 Atom CPU clocked at 1.6GHz and 1GB of RAM.
It runs on Debian 9 and I use it for inference only !!!
I have models using XGBoost and thatās fine. I had to struggle a bit more with my PyTorch models as it is 32bit device and not PyTorch is not working on 32bit systems. So I have broken my models down to play them with the simple linear algebra using numpy.
Of course for training I have other devices : a MSI laptop (i7, 16GB, GTX 1060) and a retired open air mining rig (AMD CPU, 12GB, GTX 1080) but no data science war machine.
All models are running on cloud. Start fromAzure ML studio. Moved to Colab and Python as others suggested. Moved to Kaggle notebook later and now using Deepnote to conduct daily submission.