LLMs still do not locate bounding boxes well

I sent an image to over a dozen LLMs that support vision, asking them:

Detect objects in this 1280×720 px image and return their color and bounding boxes in pixels. Respond as a JSON object: {[label]: [color, x1, y1, x2, y2], …}

None of the models did a good-enough job. It looks like we have some time to go before LLMs become good at bounding boxes.

I’ve given them a subjective rating on a 1-5 scale below.

ModelPositionsSizes
gemini-1.5-flash-001🟢🟢🟢🔴🔴🟢🟢🟢🟢🔴
gemini-1.5-flash-8b🟢🟢🟢🔴🔴🟢🟢🟢🔴🔴
gemini-1.5-flash-002🟢🟢🔴🔴🔴🟢🟢🟢🔴🔴
gemini-1.5-pro-002🟢🟢🟢🔴🔴🟢🟢🟢🟢🔴
gpt-4o-mini🟢🔴🔴🔴🔴🟢🟢🔴🔴🔴
gpt-4o🟢🟢🟢🟢🔴🟢🟢🟢🟢🔴
chatgpt-4o-latest🟢🟢🟢🟢🔴🟢🟢🟢🟢🔴
claude-3-haiku-20240307🟢🔴🔴🔴🔴🟢🟢🔴🔴🔴
claude-3=5-sonnet-20241022🟢🟢🟢🔴🔴🟢🟢🟢🔴🔴
llama-3.2-11b-vision-preview🔴🔴🔴🔴🔴🔴🔴🔴🔴🔴
llama-3.2-90b-vision-preview🟢🟢🟢🔴🔴🟢🟢🟢🔴🔴
qwen-2-vl-72b-instruct🟢🟢🟢🔴🔴🟢🟢🔴🔴🔴
pixtral-12b🟢🟢🔴🔴🔴🟢🟢🟢🔴🔴

I used an app I built for this.

Here is the original image along with the individual results.

Update

Adding gridlines with labeled axes helps the LLMs. (Thanks @Bijan Mishra.) Here are a few examples:

Leave a Comment

Your email address will not be published. Required fields are marked *