This paper investigates how large-scale, pretrained vision-language models, particularly OpenAI’s CLIP, reflect and entrench human biases concerning pets’ physical attributes and adoptability. Recognizing that CLIP is trained on vast amounts of uncurated web data and thus contains inherent biases, the authors propose using these models to explore human perceptions in human-animal relationships. Preliminary experiments using dog images and fifty dog-related phrases revealed that images of white dogs were the best match for phrases like ‘adoptable dog’ and ‘I want to adopt this dog’, raising questions about the ‘black dog syndrome’. The study also found negative correlations between “mixed-breed dog” and phrases such as “likeable dog” or “special dog”. The paper suggests these AI models can offer insights into human preferences and adoptability factors while emphasizing the need for caution, larger datasets, and better understanding of the models’ internal decision-making processes.
Non-Invasive Computer Vision-Based Fruit Fly Larvae Differentiation: Ceratitis capitata and Bactrocera zonata
This paper proposes a novel, non-invasive method using computer vision