Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Updated Training Set and New Results -- L3C v3 #14

Open
fab-jul opened this issue Mar 6, 2020 · 2 comments
Open

Updated Training Set and New Results -- L3C v3 #14

fab-jul opened this issue Mar 6, 2020 · 2 comments

Comments

@fab-jul
Copy link
Owner

fab-jul commented Mar 6, 2020

Previously, our preprocessing script saved all training images and validation images as JPGs with a high quality factor of Q=95, downscaled by a factor 0.75. It turns out that the resulting images have a specific enough distribution that the neural network picks up on it, and the images are also easier to compress for the non-learned codecs.

For correctness, we have thus re-created the training and validation sets. The new preprocessing script is available in the repo. The important differences are:

  • All images are saved as PNGs.
  • We do not rescale validation sets in any way, and instead divide the images into crops such that everything fits into memory. Note that this is a bias against our method, since more context can only help. We only crop images too big to fit into our GPU (TITAN X Pascal). Please see the updated README.
  • For the training set, we use a random downscaling factor, instead of fixed 0.75x: this provides a wider variety of downscaling artefacts.
  • Additionally, we use the Lanczos filter, as we found that Bicubic also introduces specific artefacts.

This causes all results to shift, however, as before, we still outperform WebP, JPEG-2000, and PNG, i.e. the ordering of the methods according to bpp remains unchanged.

We evaluated our model on 500 images randomly selected from the Open Images validation set, and preprocessed like the training data. To compare, please download Open Images evaluation set here.

Updated ArXiv

Available here https://arxiv.org/abs/1811.12817v3.

New Results

  Method Open Images DIV2K RAISE-1k
Ours L3C v3 2.990597132 3.093768752 2.386501087
Learned Baselines RGB Shared 4.313588005 4.429001861 3.779201962
  RGB 3.297824781 3.418117799 2.572320659
Non-Learned Approaches PNG 4.004512908 4.234729262 3.556403138
  JPEG2000 3.054759549 3.126744435 2.46459739
  WebP 3.047477818 3.176081706 2.461481317
  FLIF 2.866778476 2.910950783 2.084036243

Status

Merged into master.

@fab-jul fab-jul changed the title Updated Training Set and New Results -- L3C v2 Updated Training Set and New Results -- L3C v3 Mar 6, 2020
@fab-jul fab-jul pinned this issue Mar 6, 2020
@fab-jul fab-jul unpinned this issue Mar 6, 2020
@fab-jul
Copy link
Owner Author

fab-jul commented Mar 11, 2020

Arxiv is now available https://arxiv.org/abs/1811.12817v3.

Merged feat/v3 into master.

Please let me know here if there are issues with this release.

@DonaldTsang
Copy link

DonaldTsang commented Apr 14, 2020

Sorry for cross-posting but caoscott/SReC#6 has addressed that there are newer image formats like AVIF, HEIF and JPEG-XL (JXL).

For other specific training sets, maybe caoscott/SReC#5 can provide some tests for specific niches

@fab-jul fab-jul pinned this issue Apr 25, 2020
@fab-jul fab-jul unpinned this issue Apr 25, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants