Recently, the vision transformer (ViT) has achieved remarkable performance in computer vision tasks and has been actively utilized in colorization. Vision transformer uses multi-head self attention to effectively propagate user hints to distant relevant areas in the image. However, despite the success of vision transformers in colorizing the image, heavy underlying ViT architecture and the large c...Show More
Group-level pruning is a model compression method that can accelerate models on general hardware while maintaining the accuracy of models even at high sparsity. Existing group-level pruning studies have a limitation in that pruning is performed by evaluating the importance of the weight group based on a single global threshold, and thus the accuracy loss is significant. In this paper, we propose G...Show More