Learning pixel-adaptive weights for portrait photo retouching

Binglu Wang, Chengzhe Lu, Dawei Yan, Yongqiang Zhao, Ning Li, Xuelong Li

科研成果: 期刊稿件文章同行评审

3 引用 (Scopus)

摘要

The lookup table-based methods achieve promising retouching performance by learning image-adaptive weights to combine 3-dimensional lookup tables (3D LUTs) and conducting pixel-to-pixel color transformation. However, this paradigm ignores the local context cues and applies the same transformation to portrait pixels and background pixels that exhibit the same raw RGB values. In contrast, an expert usually conducts different operations to adjust the color temperatures, tones of portrait regions, and background regions. This inspires us to model local context cues to improve the retouching quality explicitly.Thus, the center pixel of an image patch is first retouched by predicting pixel-adaptive lookup table weights. To modulate the influence of neighboring pixels, as neighboring pixels exhibit different affinities to the center pixel, a local attention mask is estimated. Then, the quality of the local attention mask is further improved by applying supervision, which is based on the affinity map calculated by the ground-truth portrait mask. For group-level consistency, we propose to directly constrain the variance of mean color components in the Lab space. Extensive experiments on the PPR10K dataset demonstrate the effectiveness of the proposed method, the retouching performance on high-resolution photos is improved by over 0.5dB in terms of PSNR, and the group-level inconsistency is reduced by 2.1.

源语言英语
文章编号109775
期刊Pattern Recognition
143
DOI
出版状态已出版 - 11月 2023

指纹

探究 'Learning pixel-adaptive weights for portrait photo retouching' 的科研主题。它们共同构成独一无二的指纹。

引用此