Learning pixel-adaptive weights for portrait photo retouching

  • Binglu Wang
  • , Chengzhe Lu
  • , Dawei Yan
  • , Yongqiang Zhao
  • , Ning Li
  • , Xuelong Li

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

The lookup table-based methods achieve promising retouching performance by learning image-adaptive weights to combine 3-dimensional lookup tables (3D LUTs) and conducting pixel-to-pixel color transformation. However, this paradigm ignores the local context cues and applies the same transformation to portrait pixels and background pixels that exhibit the same raw RGB values. In contrast, an expert usually conducts different operations to adjust the color temperatures, tones of portrait regions, and background regions. This inspires us to model local context cues to improve the retouching quality explicitly.Thus, the center pixel of an image patch is first retouched by predicting pixel-adaptive lookup table weights. To modulate the influence of neighboring pixels, as neighboring pixels exhibit different affinities to the center pixel, a local attention mask is estimated. Then, the quality of the local attention mask is further improved by applying supervision, which is based on the affinity map calculated by the ground-truth portrait mask. For group-level consistency, we propose to directly constrain the variance of mean color components in the Lab space. Extensive experiments on the PPR10K dataset demonstrate the effectiveness of the proposed method, the retouching performance on high-resolution photos is improved by over 0.5dB in terms of PSNR, and the group-level inconsistency is reduced by 2.1.

Original languageEnglish
Article number109775
JournalPattern Recognition
Volume143
DOIs
StatePublished - Nov 2023

Keywords

  • 3D Lookup table
  • Portrait photo retouching
  • Visual attention

Fingerprint

Dive into the research topics of 'Learning pixel-adaptive weights for portrait photo retouching'. Together they form a unique fingerprint.

Cite this