文摘
This paper addresses the topic of image colorization that consists in converting a gray-scale image into a color one. In the literature, there exist two main types of approaches to tackle this problem. The first one is the manual methods where the color information is given by some scribbles drawn by the user on the image. The interest of these approaches comes from the interactions with the user that can put any color he wants. Nevertheless, when the scene is complex many scribbles must be drawn and the interactive process becomes tedious and time-consuming. The second category of approaches is the exemplar-based methods that require a color image as input. Once the example image is given, the colorization is generally fully automatic. A limitation of these methods is that the example image needs to contain all the desired colors in the final result. In this paper, we propose a new framework that unifies these two categories of approaches into a joint variational model. Our approach is able to take into account information coming from any colorization method among these two categories. Experiments and comparisons demonstrate that the proposed approach provides competitive colorization results compared to state-of-the-art methods.