ThumbNet: One Thumbnail Image Contains All You Need for Recognition
Abstract
Although deep convolutional neural networks (CNNs) have achieved great success in computer vision tasks, its real-world application is still impeded by its voracious demand of <PRE_TAG>computational resources</POST_TAG>. Current works mostly seek to compress the network by reducing its parameters or parameter-incurred computation, neglecting the influence of the input image on the system complexity. Based on the fact that input images of a CNN contain substantial redundancy, in this paper, we propose a unified framework, dubbed as ThumbNet, to simultaneously accelerate and compress CNN models by enabling them to infer on one thumbnail image. We provide three effective strategies to train ThumbNet. In doing so, ThumbNet learns an inference network that performs equally well on small images as the original-input network on large images. With ThumbNet, not only do we obtain the <PRE_TAG>thumbnail-input <PRE_TAG>inference network</POST_TAG></POST_TAG> that can drastically reduce computation and memory requirements, but also we obtain an image downscaler that can generate thumbnail images for generic classification tasks. Extensive experiments show the effectiveness of ThumbNet, and demonstrate that the <PRE_TAG>thumbnail-input <PRE_TAG>inference network</POST_TAG></POST_TAG> learned by ThumbNet can adequately retain the accuracy of the original-input network even when the input images are downscaled 16 times.
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper