great overview, nice to see Xception is mentioned as an emerging backbone network - this is what i'm working with right nowhttps://twitter.com/graphific/status/1041627931842420742 …
You can add location information to your Tweets, such as your city or precise location, from the web and via third-party applications. You always have the option to delete your Tweet location history. Learn more
Thats means that all CNNs have a minimum input size, which is the minimum size you can call the network on without resulting in empty feature maps. At that size, the output of your network is a 1x1 feature map.
For most nets, this tends to be about ~50% of the input size the network was designed for. Downsampling is absolutely necessary when you have large inputs. To get rid of this limitation, you need to create your own network, where you would do less downsampling.
That's only partially true: for discriminative models yes, but there's no inherent reason why convolutions should equals downsampling (for generative models for instance).
*Of course* you can design architectures that don't do downsampling (just don't include maxpooling or strides, and use "same" padding everywhere). And there are use cases for it. I'm just talking about every pre-trained architecture available out there.
sure, i understand that - my data is 180X180 and i'm using Xception and it works fine
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.