V* - Better than GPT-4V? Iterative Context Refining for Visual Question Answer!
Is V* really better than GPT-4V at Visual Question Answering (VQA)?
V* is a way to augment the prompt for Visual Question Answer (VQA) to be more than just the image and the question itself, but also a list of target objects that can help with the question and their positions.
This list of target objects can be found via a Visual Search Model. This Visual Search Model starts off with the full image and tries to find the target object's bounding boxes. If unable to find, it uses the heatmap which matches a contextual cue to the target object, and identifies a quadrant of the original image which the target object can be found in. The process then continues with this quadrant until we reach the minimum image size.
This iterative focusing of the image helps to mitigate the lack of positional sensitivity of the Vision Transformer embeddings for the image encoding of the multimodal Large Language Model (LLM).
In general, this approach of adding relevant context and searching the image by focusing on the right sub-sections is a very powerful one. I also show that if we incorporate some aspects of the V* method into GPT-4V, it can help improve the performance of GPT-4V!
~~~
References:
V* Github: https://vstar-seal.github.io/
My Slides: https://github.com/tanchongmin/TensorFlow-Implementations/blob/main/Paper_Reviews/Vstar.pdf
Vision Transformer: https://arxiv.org/abs/2010.11929
CLIP embeddings: https://arxiv.org/abs/2103.00020
LLaVA: Large Language and Vision Assistant: https://arxiv.org/abs/2304.08485
GPT-4V Technical Report: https://arxiv.org/abs/2303.08774
Chain-of-thought: https://arxiv.org/abs/2201.11903
ReAct framework: https://arxiv.org/abs/2210.03629
~~~
0:00 Introduction
2:13 Key issue with CLIP embeddings based on ViT
13:31 Background: LlaVA
18:05 Overall Walkthrough of V*
31:12 Visual Search Model
39:52 Visual QA Model
43:33 Iterative Visual Search
45:05 V* Example 1
52:00 V* Example 2
59:26 A form of Best First Search
1:02:42 How to improve V* (Great discussion with Richard)
1:10:46 Putting Everything Together
1:13:38 Comparison: Chain of Thought
1:15:17 Comparison: ReAct Framework
1:16:53 Results
1:22:11 My experiments: Incorporating V* into GPT-4V
1:23:32 V* is actually less generic than GPT-4V
1:24:33 V* heuristic-based search based on heat map is similar to human fixation!
1:26:10 My takeaways
1:26:47 Discussion and Conclusion
~~~~
AI and ML enthusiast. Likes to think about the essences behind breakthroughs of AI and explain it in a simple and relatable way. Also, I am an avid game creator.
Discord: https://discord.gg/bzp87AHJy5
LinkedIn: https://www.linkedin.com/in/chong-min-tan-94652288/
Online AI blog: https://delvingintotech.wordpress.com/
Twitter: https://twitter.com/johntanchongmin
Try out my games here: https://simmer.io/@chongmin