1/ Higher Top 1 accuracy in image classification actually isn't a great indicator that a new vision model is going to work out well in practice in an image to image task (like DeOldify). I've learned this the hard way after getting excited about a new shiny model many times!
-
Show this thread
-
2/ EfficientNets come to mind- I just have not been seeing the benefits relative to the classification performance. But this really applies generally, even within a set of architectures that have been trained the same way (FaceBook's wsl models, BiTM, etc).
1 reply 0 retweets 8 likesShow this thread -
3/ In those cases, you just don't really know until you actually try. You may be surprised, I can tell you that much! So don't just pick the one with the highest Top-1 accuracy in ImageNet...
5 replies 1 retweet 11 likesShow this thread -
Replying to @citnaj
Any better metrics that you might not have come up with along the way? Or sets of metrics?
3 replies 0 retweets 3 likes -
Replying to @MaxLenormand
2/ ... learning tasks. Specifically- the larger and more varied the training, the better. This is basically the Facebook WSL models' secret ingredient too. I think everything else is basically noise that you have to sift through with experiments.
1 reply 0 retweets 2 likes -
Replying to @citnaj @MaxLenormand
Justin Retweeted Justin
A recent paper by Karras also seems to suggest that for transfer learning of StyleGAN models diversity of training data is really important too.https://twitter.com/Buntworthy/status/1271554286917468161 …
Justin added,
1 reply 0 retweets 2 likes -
Is there an established way to measure how "varied" a training dataset is?
1 reply 0 retweets 2 likes
That's a great question. I honestly don't know off the top of my head...
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.