Published January 1, 2023 | Version v1
Conference paper Open

Relative Attributes Classification via Transformers and Rank SVM Loss

  • 1. Univ Surrey, CVSSP, Guildford, Surrey, England
  • 2. Sabanci Univ, Ctr Excellence Data Analyt VERIM, Istanbul, Turkiye

Description

We propose a new model for learning to rank two images with respect to their relative strength of expression for a given attribute. We address this problem - called relative attribute learning - using a vision transformer backbone. The embedded representations of the two images to be compared are extracted and used for comparison with a ranking head, in an end-to-end fashion. The results demonstrate the strength of vision transformers and their suitability for relative attributes classification. Our proposed approach outperforms the state-of-the-art by a large margin, achieving 90.40% and 98.14% mean accuracy over the attributes of LFW-10 and Pubfig datasets.

Files

bib-8d3ded40-f2f7-4f64-ae50-6088038323ec.txt

Files (170 Bytes)

Name Size Download all
md5:478fa76cb12ea9fe0192745bf4bc7d08
170 Bytes Preview Download