作者: Ian D. Reid , Chunhua Shen , Anton van den Hengel , Bohan Zhuang , Qi Wu
DOI:
关键词:
摘要: Visual relationship detection aims to capture interactions between pairs of objects in images. Relationships and humans represent a particularly important subset this problem, with implications for challenges such as understanding human behaviour, identifying affordances, amongst others. In addressing problem we first construct large-scale human-centric visual dataset (HCVRD), which provides many more types annotation (nearly 10K categories) than the previous released datasets. This large label space better reflects reality human-object interactions, but gives rise long-tail distribution turn demands zero-shot approach labels appearing only test set. This is time issue has been addressed. We propose webly-supervised these problems demonstrate that proposed model strong baseline on our HCVRD dataset.