What is anisotropic scaling? And how is it achieved w.r.t image processing and computer vision?
I understand that it is some form of non-uniform scaling
as Wikipedia puts it, but I still don't get a good grasp of what it actually means when applied to images. Recently, some deep learning architectures like R-CNN for object detection also uses it but doesn't show much light on this topic.
Any examples and visual illustrations that explains the concept clearly would be really nice.
Non-uniform scaling just means that different scales are applied to each dimension, making it anisotropic. The opposite would be isotropic scaling, where the same scale is applied to each dimension.
In the context of R-CNN, the authors did a trick to use variable sized images with the classification network, they resize any image into a fixed size (I believe it was 224x224), and this is done independently of the aspect ratio of the image. So say you have a 1280x720 image, and if it is resized to 224x244, then the scales are 1280/224 and 720/224, which obviously are not the same, hence anisotropic scaling.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With