Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2023 Jan 23;13:1286. doi: 10.1038/s41598-022-26364-y

Publisher Correction: Contrastive language and vision learning of general fashion concepts

Patrick John Chia 1,, Giuseppe Attanasio 2, Federico Bianchi 3, Silvia Terragni 4,7, Ana Rita Magalhães 5, Diogo Goncalves 5, Ciro Greco 6, Jacopo Tagliabue
PMCID: PMC9870952  PMID: 36690677

Correction to: Scientific Reports https://doi.org/10.1038/s41598-022-23052-9, published online 08 November 2022

The original version of this Article contained errors in Figures 2 and 7 where the images did not display correctly. The original Figures 2 and 7 and accompanying legends appear below.

Figure 2.

Figure 2

Schematic overview of multi-modal retrieval (left) and zero-shot classification tasks (right).

Figure 7.

Figure 7

Item bounding-box detection. Localization maps can be easily extended to provide zero-shot bounding boxes for items of interest. Green bounding boxes show the predicted locations for fashion concepts “Backpack” (left) and “Straw hat” (right). Images above are taken from the publicly available Unsplash Lite Dataset 1.2.0: FashionCLIP was tested extensively on ModaNet - please reach out to authors for links to those images.

The original Article has been corrected.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES