PyNET-Q×Q: An Efficient PyNET Variant for Q×Q Bayer Pattern Demosaicing in CMOS Image Sensors
Deep learning-based image signal processor (ISP) models for mobile cameras can generate high-quality images that rival those of professional DSLR cameras. However, their computational demands often make them unsuitable for mobile settings. Additionally, modern mobile cameras employ non-Bayer color f...
Saved in:
| Main Authors: | , , , , , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2023-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/10114931/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Summary: | Deep learning-based image signal processor (ISP) models for mobile cameras can generate high-quality images that rival those of professional DSLR cameras. However, their computational demands often make them unsuitable for mobile settings. Additionally, modern mobile cameras employ non-Bayer color filter arrays (CFA) such as Quad Bayer, Nona Bayer, and <inline-formula> <tex-math notation="LaTeX">$\text{Q}\times \text{Q}$ </tex-math></inline-formula> Bayer to enhance image quality, yet most existing deep learning-based ISP (or demosaicing) models focus primarily on standard Bayer CFAs. In this study, we present PyNET-<inline-formula> <tex-math notation="LaTeX">$\text{Q}\times \text{Q}$ </tex-math></inline-formula>, a lightweight demosaicing model specifically designed for <inline-formula> <tex-math notation="LaTeX">$\text{Q}\times \text{Q}$ </tex-math></inline-formula> Bayer CFA patterns, which is derived from the original PyNET. We also propose a knowledge distillation method called progressive distillation to train the reduced network more effectively. Consequently, PyNET-<inline-formula> <tex-math notation="LaTeX">$\text{Q}\times \text{Q}$ </tex-math></inline-formula> contains less than 2.5% of the parameters of the original PyNET while preserving its performance. Experiments using <inline-formula> <tex-math notation="LaTeX">$\text{Q}\times \text{Q}$ </tex-math></inline-formula> images captured by a prototype <inline-formula> <tex-math notation="LaTeX">$\text{Q}\times \text{Q}$ </tex-math></inline-formula> camera sensor show that PyNET-<inline-formula> <tex-math notation="LaTeX">$\text{Q}\times \text{Q}$ </tex-math></inline-formula> outperforms existing conventional algorithms in terms of texture and edge reconstruction, despite its significantly reduced parameter count. Code and partial datasets can be found at <uri>https://github.com/Minhyeok01/PyNET-QxQ</uri>. |
|---|---|
| ISSN: | 2169-3536 |