RUS  ENG
Полная версия
ЖУРНАЛЫ // Компьютерная оптика // Архив

Компьютерная оптика, 2021, том 45, выпуск 4, страницы 615–626 (Mi co947)

Эта публикация цитируется в 4 статьях

ЧИСЛЕННЫЕ МЕТОДЫ И АНАЛИЗ ДАННЫХ

Veiling glare removal: synthetic dataset generation, metrics and neural network architecture

A. V. Shoshinab, E. A. Shvetsb

a Moscow Institute of Physics and Technology (State University), Institutsky per. 9, Dolgoprudny, 141701, Russia
b Kharkevich Institute for Information Transmission Problems, RAS, Bolshoy Karetny per. 19, build.1, Moscow, 127051, Russia

Аннотация: In photography, the presence of a bright light source often reduces the quality and readability of the resulting image. Light rays reflect and bounce off camera elements, sensor or diaphragm causing unwanted artifacts. These artifacts are generally known as "lens flare" and may have different influences on the photo: reduce contrast of the image (veiling glare), add circular or circular-like effects (ghosting flare), appear as bright rays spreading from light source (starburst pattern), or cause aberrations. All these effects are generally undesirable, as they reduce legibility and aesthetics of the image. In this paper we address the problem of removing or reducing the effect of veiling glare on the image. There are no available large-scale datasets for this problem and no established metrics, so we start by (i) proposing a simple and fast algorithm of generating synthetic veiling glare images necessary for training and (ii) studying metrics used in related image enhancement tasks (dehazing and underwater image enhancement). We select three such no-reference metrics (UCIQE, UIQM and CCF) and show that their improvement indicates better veil removal. Finally, we experiment on neural network architectures and propose a two-branched architecture and a training procedure utilizing structural similarity measure.

Ключевые слова: lens flare, veiling glare, image enhancement, deep learning, synthetic data.

Поступила в редакцию: 26.02.2021
Принята в печать: 28.04.2021

Язык публикации: английский

DOI: 10.18287/2412-6179-CO-883



© МИАН, 2024