2023 Data Compression Conference (DCC) 2023
DOI: 10.1109/dcc55655.2023.00012
|View full text |Cite
|
Sign up to set email alerts
|

Learned Disentangled Latent Representations for Scalable Image Coding for Humans and Machines

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 10 publications
0
6
0
Order By: Relevance
“…For the YOLOv3 model, we test our models on a subset of 5000 images from the COCO2014 [24] dataset, and use the mean average precision (mAP) as measured at 50% intersection over union (IoU), which is denoted as mAP@50, as our accuracy metric. Our results for YOLOv3 are compared with the best published settings of 3 previous scalable codecs [2,3,4], refered to as Choi2022, Harell2022, and Ozyilkan2023. For completeness, we also include two traditional codecs, VVCintra [16] and HEVC-intra [14] (also known as BPG), and the learnable codec of [10] to which we refer to as Cheng2020 3 .…”
Section: Base Layer Resultsmentioning
confidence: 99%
See 4 more Smart Citations
“…For the YOLOv3 model, we test our models on a subset of 5000 images from the COCO2014 [24] dataset, and use the mean average precision (mAP) as measured at 50% intersection over union (IoU), which is denoted as mAP@50, as our accuracy metric. Our results for YOLOv3 are compared with the best published settings of 3 previous scalable codecs [2,3,4], refered to as Choi2022, Harell2022, and Ozyilkan2023. For completeness, we also include two traditional codecs, VVCintra [16] and HEVC-intra [14] (also known as BPG), and the learnable codec of [10] to which we refer to as Cheng2020 3 .…”
Section: Base Layer Resultsmentioning
confidence: 99%
“…For Faster R-CNN, we use the entire COCO2017 [24] validation set (which also contains 5000 images), and report the average mAP over a range of IoU thresholds between 50 − 95% with steps of 5%, which we simply denote mAP 4 . Available benchmarks for comparison here are Choi2022, Cheng2020 and the two traditional codecs VVC and HEVC.…”
Section: Base Layer Resultsmentioning
confidence: 99%
See 3 more Smart Citations