Hello!
I used the example of GpuKMeans to write a KMeans application over MNIST dataset that contains 60000 images, on the other hand, I used org.apache.spark.ml.clustering API over the same dataset in order to compare performance...
this is what I got as results for 20 iterations:
CPU: 16.54 s
GPU: 41,1 s
It's 2.48 slowdown !!!*
So how can I make KMeans application run faster on GPU ?
Ps: those results are obtained on spark-shell fired over 40 cores and M60 GPU
Best regards
aguerzaa