Evaluation

The SpaRe Benchmark provides a comprehensive evaluation of the submitted solutions.

Metrics

Submissions are evaluated based on the following metrics:

  • Masked PSNR (PSNR-M) - PSNR calculated with the object mask,
  • PSNR - calculated in the whole image, including background reconstruction,
  • SSIM-M - SSIM calculated withing the tight bounding box around the object,
  • SSIM - calculated in the whole image,
  • LPIPS-M - LPIPS calculated withing the tight bounding box around the object, using AlexNet as a backbone,
  • LPIPS - calculated in the whole image.

Data

All submissions are evaluated on SpaRe Dataset (Synthetic), and DTU in agreement with [1].

The benchmark contains data SpaRe Dataset and DTU, Codabench allows submission of solution including either of those datasets or both.

Users are provided with all metrics in the following details:

  • separately for SpaRe Dataset and DTU (only one can be submitted to the benchmark),
  • on various granularity level: for each image, average for every scene, average for the whole dataset.

[1] Nazarczuk, M., Tanay, T., Catley-Chandar, S., Shaw, R., Timofte, R., PĂ©rez-Pellitero, E.: AIM 2024 Sparse Neural Rendering Challenge: Dataset and Benchmark. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2024)