참고
[1] Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, Armand Joulin, “Unsupervised learning of visual features by contrasting cluster assignments”, NeurIPS 2020.

[2] Ting Chen, Simon Kornblith, Mohammad Norouzi, Geoffrey Hinton, “A simple framework for contrastive learning of visual representations”, ICML 2020.

[3] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, Li Fei-Fei, “ImageNet: a large-scale hierarchical image database”, CVPR 2009.

[4] Jean-Bastien Grill, Florian Strub, Florent Altche, Corentin Tallec, Pierre H. Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, Bilal Piot, Koray Kavukcuoglu, Remi Munos, Michal Valko, “Bootstrap your own latent: a new approach to self-supervised learning”, NeurIPS 2020.

[5] Geoffrey Hinton, Oriol Vinyals, Jeff Dean, “Distilling the knowledge in a neural network”, arXiv:1503.02531, 2015.

[6] Gabriel Ilharco, et al., OpenCLIP, 2021, https://github.com/mlfoundations/open_clip.

[7] Byoungjip Kim, Sungik Choi, Dasol Hwang, Moontae Lee, Honglak Lee, “Transferring pre-trained multimodal representations with cross-modal similarity matching”, NeurIPS 2022.

[8] Maria-Elena Nilsback and Andrew Zisserman, “Automated flower classification over a large number of classes”, ICCVGIP 2008.

[9] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever, “Learning transferable visual models from natural language supervision”, arXiv:2103.00020, 2021.

[10] Yonglong Tian, Dilip Krishnan, Phillip Isola, “Contrastive representation distillation”, ICLR 2020.