A New Data Source and Learning Paradigm for Multimodal LLMs (2024)

An Yan,Zhengyuan Yang,Junda Wu,Wanrong Zhu,Jianwei Yang,Linjie Li,
Kevin Lin,Jianfeng Wang,Julian McAuley,Jianfeng Gao,Lijuan Wang
UC San DiegoMicrosoft CorporationUC Santa Barbara
{ayan,juw069,jmcauley}@ucsd.edu, wanrongzhu@ucsb.edu,
{zhengyang,jianwei.yang,keli,lindsey.li,jianfw,jfgao,lijuanw}@microsoft.com

Abstract

Set-of-Mark (SoM) Prompting unleashes the visual grounding capability of GPT-4V, by enabling the model to associate visual objects with tags inserted on the image. These tags, marked with alphanumerics, can be indexed via text tokens for easy reference. Despite the extraordinary performance from GPT-4V, we observe that other Multimodal Large Language Models (MLLMs) struggle to understand these visual tags. To promote the learning of SoM prompting for open-source models, we propose a new learning paradigm: “list items one by one,” which asks the model to enumerate and describe all visual tags placed on the image following the alphanumeric orders of tags. By integrating our curated dataset with other visual instruction tuning datasets, we are able to equip existing MLLMs with the SoM prompting ability. Furthermore, we evaluate our finetuned SoM models on five MLLM benchmarks. We find that this new dataset, even in a relatively small size (10k-30k images with tags), significantly enhances visual reasoning capabilities and reduces hallucinations for MLLMs. Perhaps surprisingly, these improvements persist even when the visual tags are omitted from input images during inference. This suggests the potential of “list items one by one” as a new paradigm for training MLLMs, which strengthens the object-text alignment through the use of visual tags in the training stage. Finally, we conduct analyses by probing trained models to understand the working mechanism of SoM. Our code and data are available at https://github.com/zzxslp/SoM-LLaVA.

1 Introduction

Recent advances in Multimodal Large Language Models (MLLMs) such as GPT-4V(OpenAI, 2023a) show strong performance in multimodal perception and reasoning, enabling various new capabilities(Yang etal., 2023b). Among these, Set-of-Mark Prompting (SoM)(Yang etal., 2023a) is an interesting new working mode that enhances the connection between visual objects and textual tokens via visual prompting, i.e., placing alphanumeric tags on input images. It provides a natural interface for human-computer interaction, by linking visual locations to executable actions through visual tags, and enables various applications such as GUI navigation(Yan etal., 2023b) and robot interaction(Lin etal., 2023a). Furthermore, GPT-4V with SoM(Yang etal., 2023a) can implicitly align visual objects with their corresponding tags. Such alignments(Li etal., 2020; Yang etal., 2021) allow MLLMs to leverage index numbers to perform multi-hop visual reasoning(Yang etal., 2023a; Wei etal., 2022), thereby improving their abilities in multimodal understanding and reasoning tasks.

A New Data Source and Learning Paradigm for Multimodal LLMs (1)

Despite the significant interest in SoM prompting and its broad applications, it remains unclear why GPT-4V can benefit from SoM prompting,We find that other MLLMs, including the state-of-the-art open-sourced models such as LLaVA-v1.5(Liu etal., 2024), and commercial systems like Gemini(Team etal., 2023), struggle to understand SoM prompts.This gap prevents them from leveraging the effectiveness of SoM prompting.In this study, we aim to deepen the understanding of SoM, with a goal of facilitating arbitrary MLLMs to benefit from it.

We break down SoM prompting into three core capabilities: (1) the ability to identify all tags and read the alphanumeric scene texts written on them;(2) the ability to recognize and pinpoint all objects in an image;(3) the ability to associate tags with corresponding objects in the image.Despite possessing skills such as OCR and visual recognition to meet the first two capabilities, most MLLMs still fail to fully understand SoM prompts.Therefore, we hypothesize that the crucial missing element is the third capability, associating tags with objects, which requires deliberate training.We further validate that SoM-style data are sparse in common MLLM training sources, and it may be necessary to create a specific dataset.

To facilitate such training, we introduce a new learning paradigm named “list items one by one”. We show that by asking MLLMs to comprehensively list all tagged items following the alphanumeric order of visual tags, MLLMs can learn SoM prompting with a small number of item-listing samples. Specifically, we create a tailored dataset, by tagging images with Semantic-SAM(Li etal., 2023c; Yang etal., 2023a), and prompting GPT-4V to generate paired text descriptions. With just 10k image-text pairs, MLLMs like LLaVA-1.5(Liu etal., 2023a) can reliably understand SoM tags. Based on this initial finding, we conduct studies to explore the effective recipes to help MLLMs best utilize SoM prompting.

We enhanced MLLMs with this “list items one by one” objective and assess their SoM performance from two aspects:model’s ability to recognize and describe the SoM tags, and its ability to use SoM in improving multimodal reasoning (Figure1). For the first aspect, we design the tag listing task, which requires MLLMs to list and describe all tags in the image, evaluated by listing accuracy. For the second aspect, we evaluate finetuned models on five MLLM benchmarks, including POPE, MME, SEED-Bench, LLaVA-Bench, and MM-Vet, showcasing that MLLMs with SoM can significantly boost the multmodal understanding performance. Moreover, our model trained with SoM data outperforms the original MLLM, even without additional visual tags during inference.This demonstrates the potential of incorporating our proposed dataset and learning paradigm to boost general MLLM training.

Finally, we revisit our original question regarding the working mechanism of SoM. The preliminary hypothesis is that the SoM capability may be related to OCR and the implicit association among text, tags, and objects. With our trained models, specifically SoM-LLaVA, we gain access to model features and attention maps for an in-depth analysis. We visualize the attention map to verify tag association. Compared with the original LLaVA model, SoM-LLaVA indeed learns better visual-tag-text associations, reflected in corresponding attention maps.

Our contributions are summarized as follows.

  • We present a new training task and data source named “list items one by one,” which effectively bootstraps MLLMs for the SoM visual prompting ability.

  • We evaluate our finetuned SoM MLLMs on five multimodal understanding benchmarks, and show improved performance even when SoM tags are removed from the input image.

  • We probe the working mechanism of SoM through the trained MLLMs, showcasing the implicit association between visual objects and text tokens when performing SoM prompting.

2 Related Work

Visual referring prompting.

Other than text prompts, visual referring prompting(Yang etal., 2023b) is another effective approach when interacting with multimodal LLMs, where users directly draw on input images to specify their intent, such as drawing visual pointers or handwriting scene texts. Early studies show that vision-language models can understand visual pointers such as circles(Shtedritski etal., 2023) and dots(Mani etal., 2020). Recent studies(Yang etal., 2023b)show that more powerful multimodal LLMs(OpenAI, 2023a) can handle more complicated prompts such as arrows, boxes, circles, hand drawing, scene text, as well as their combinations. Another major advancement is Set-of-Mark Prompting (SoM)(Yang etal., 2023a), where numbered tags can be placed on images to associate visual objects with text indexed. Its effective visual grounding capability(Kazemzadeh etal., 2014; Yu etal., 2016; Mao etal., 2016) enables various applications(Yan etal., 2023b; Zhang etal., 2023). In this work, we aim to better understand SoM and extend its success from GPT-4V(OpenAI, 2023a) to other open-source multimodal LLMs.

Multimodal LLMs.

Multimodal LLMs(Alayrac etal., 2022; Zhu etal., 2022; OpenAI, 2023a; Liu etal., 2023b; Li etal., 2023b) extend large language models(OpenAI, 2023b; Gao etal., 2023; Touvron etal., 2023) with visual perception capabilities. Recent studies(Chen etal., 2023) show the effectiveness of training open-source models on the GPT-4V generated detailed description data. Another thread of studies explore having multimodal LLMs predicting object locations as bounding boxes(Wang etal., 2023b; Peng etal., 2023) or masks(Rasheed etal., 2023).In contrast to most prior studies that pair the images with different text instructions, our study explores a new direction of how visual prompts such as SoM can improve multimodal LLMs. Specifically, we show that the SoM visual tags provide fine-grained alignments between visual objects and text tokens, thereby improving various visual reasoning tasks, both with and without SoM prompting during inference.

3 Preliminary Examination

3.1 Visualizing SoM Prompting on LLaVA

In this section, we first investigate the capacity of LLaVA-1.5 in SoM, concerning its attention sensibility to the numeric IDs tagged on the objects and its answer to the SoM query.We show an example task to list a series of objects tagged with numeric IDs in Figure 2, in which the attention map is extracted from LLaVA-1.5 based on the SoM query(e.g., “I have labeled a bright numeric ID at the center for each visual object in the image. Please enumerate their names.”).The top 20 image patches with the highest average attention weights across the user query tokens are highlighted in transparent red regions.

A New Data Source and Learning Paradigm for Multimodal LLMs (2)

We can observe from the highly attended regions of LLaVA-1.5 that the numeric ID tags can be easily and correctly attended by LLaVA-1.5 along with their associated objects (e.g., bird, vase, and branches).Such capacities in locating numeric ID tags may have been acquired by LLaVA-1.5 from its pretraining tasks in OCR and also benefited from the strong OCR abilities of the ViT feature encoder(Radford etal., 2021) adopted by LLaVA-v1.5.However, the response prompted by the user query in the first example of Figure 2 suggests that LLaVA-1.5 cannot follow the SoM instruction to list all the items.Instead of providing the object descriptions corresponding to all the numeric ID tags, LLaVA-1.5 responds with a general image caption,due to a large portion of image captioning samples in its pretraining stage.From the second example of Figure 2, we can also observe that although LLaVA-1.5 generates a list of tag IDs with object names, it cannot accurately associate the tags to corresponding objects, causing the model to hallucinate the descriptions of these objects.

3.2 Finding SoM Data in Existing Training Sources

We further look into the pretraining/instruction-tuning (IT) dataset, aiming to inspect if there are text contents with listings, or images with SOM annotations.We examine the pretraining dataset of LLaVA-v1 and v1.5(Liu etal., 2023b; a), and the IT dataset used by LLaVA-v1.5, ShareGPT4V(Chen etal., 2023), and CogVLM(Wang etal., 2023a).

Table1 shows the source of text in each dataset and the percentage of text content with a listing format.The text in the two pretraining datasets for LLaVA are image captions (either the raw caption or generated by BLIP(Dai etal., 2023)), and we did not find any text with listings in them using our parser.Aside from image captions, the IT dataset also contains instructions related to other visual tasks such as VQA. We noticed that the answers provided by GPT-4(V) models sometimes construct the text in a listing manner (e.g., list out possible reasons for a question, list out observed objects in the image, etc). More examples can be found in AppendixA.6. The instruction-following dataset used by CogVLM has the highest percentage of text with listings (similar-to\sim7%).Through our interaction with these models, we also find CogVLM is better at generating listing-style data than LLaVA-1.5.

We add tags to MSCOCO-2017 images following the SoM(Yang etal., 2023a) format, and train a binary classifier with ViT/B-16(Dosovitskiy etal., 2020). We use the classifiers to filter the images in the two LLaVA pretraining datasets, and take the top 2k images with the highest scores for each dataset. We then manually check the top 2k images, and found 12 images with tagging in CC3M-595K (similar-to\sim0.002%), and found 86 images with tagging in LCS-558K (similar-to\sim0.015%).Figure15 shows a few images with tagging.Given that tagged images are sparse in those datasets and the SoM prompting performance of open-source MLLMs is unsatisfying, it may be worthwhile to design a tailored dataset that empower open-source MLLMs with this emergent ability, similar to what GPT-4V is capable of.

#Dataset#TextText w/ ListingSource of Text1LLaVA-Pretrain-CC3M-595K595.4K0Raw CC3M image captions.2LLaVA-Pretrain-LCS-558K558.1K0Captioned by BLIP.3LLaVA-v1.5-Mix665K3356.2K0.72%Rule-based, or generated by ShareGPT or GPT4-0314.4ShareGPT4V102.0K0.21%Generated by GPT4-Vision.5CogVLM333.5K7.16%Generated by MiniGPT4 or by GPT4-0314.

4 Dataset Creation and Training

Motivated bythe above analysis,in this section, we introduce the pipeline to create our dataset. First, inSection4.1, we use semantic-SAM to generate semantic visual prompts in the form of numeric tags for each image. We then discuss the learning paradigm of “list items one by one” inSection4.2. Finally, we use visual prompted images to generate text data in Section4.3.

4.1 Image Source and Visual Prompting Generation

There are various open-source image datasets available(Deng etal., 2009; Lin etal., 2014; Schuhmann etal., 2022; Yan etal., 2023a).We use MS-COCO(Lin etal., 2014) as the image source to create our SoM dataset, since it contains comprehensive human annotations with bounding boxes, masks, and captions. It has also been widely used for visual instruction tuning(Liu etal., 2023b; Wang etal., 2023a; Chen etal., 2023), which could benefit controlled experiments as well as comparisons with previous work.

The first step is to create visual prompts by placing numeric tags on proper locations. Following SoM(Yang etal., 2023a), we experiment with segmentation models including SEEM(Zou etal., 2023), Semantic-SAM(Li etal., 2023c), and SAM(Kirillov etal., 2023). Empirically, we find that Semantic-SAMprovides the annotation granularity that best fits COCO images, and thus use it to create tagged images for our dataset.

4.2 A Learning Paradigm: List Items One by One

After obtaining the image data with semantic tags, the next question is how to design the instruction data to best distill the SoM visual prompting ability.A common approach(Liu etal., 2023b; Chen etal., 2023) in multimodal instruction-following data creation is to design and collect “question-answering” style samples. This is often done by prompting ChatGPT/GPT-4 or alternative open-source models. Given an image I𝐼Iitalic_I and optional metadata MIsubscript𝑀𝐼M_{I}italic_M start_POSTSUBSCRIPT italic_I end_POSTSUBSCRIPT such as captions, bounding boxes, various questions or instructions XQ(i)superscriptsubscript𝑋Q𝑖X_{\texttt{Q}}^{(i)}italic_X start_POSTSUBSCRIPT Q end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT are posed, and the corresponding answers XA(i)superscriptsubscript𝑋A𝑖X_{\texttt{A}}^{(i)}italic_X start_POSTSUBSCRIPT A end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT from large models are collected.

However, such general question-answering data may not be the most effective in distilling the desired SoM prompting capability, due to the inadequate mention of objects in text. For SoM prompting, one core ability of interest is to associate numbered tags with visual objects in the image, thereby enabling effective referral of visual objects via text tokens. In a general QA data, however, it is rare for multiple objects to be mentioned, even in an extended multi-turn conversation.To enhance tag association, we propose a simple and effective approach: list items one by one, where the model is asked to comprehensively describe all tagged items within an image.Given an image ITsuperscript𝐼TI^{\texttt{T}}italic_I start_POSTSUPERSCRIPT T end_POSTSUPERSCRIPT with N𝑁Nitalic_N text tags on the image, we ask the model to enumerate all items in numerical order: {Xobj1superscriptsubscript𝑋𝑜𝑏𝑗1X_{obj}^{1}italic_X start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT, Xobj2superscriptsubscript𝑋𝑜𝑏𝑗2X_{obj}^{2}italic_X start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, \cdots, XobjNsuperscriptsubscript𝑋𝑜𝑏𝑗𝑁X_{obj}^{N}italic_X start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT}, where Xobjjsuperscriptsubscript𝑋𝑜𝑏𝑗𝑗X_{obj}^{j}italic_X start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT is the textual description of the j𝑗jitalic_j-th item, tagged by ID j𝑗jitalic_j in the image.

Beyond promoting SoM learning, listing items one by one is also effective in general multi-modal LLM training:if a model learns to list items in the images with a specific order (in our case, the order is determined by the visual numeric tags), it gains a comprehensive and fine-grained understanding of images. This could directly benefit visual grounding and reasoning, which we verified through the standard multimodal QA and chat evaluation benchmarks.

Compared with existing visual instruction tuning datasets, such as LLaVA-665K(Liu etal., 2023a) and ShareGPT-4V(Chen etal., 2023), another difference is the implicit spatial information encoded by the visual tags in SoM prompting.Converting images into the language space inevitably loses information, especially spatial locations. For example, “a girl on the right” can only vaguely imply the position of the girl. However, with SoM visual prompting, we provide precise visual guidance on the image. Therefore, our data can be viewed as a form of dense captioning with a new way of encoding spatial information.

4.3 Text Data Generation via GPT-4V

With the visual prompting enhanced images,the final step for dataset creation is to generate the corresponding text data.To automate this process, we leverage GPT-4V(OpenAI, 2023a) to generate the listing data {Xobj1superscriptsubscript𝑋𝑜𝑏𝑗1X_{obj}^{1}italic_X start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT, Xobj2superscriptsubscript𝑋𝑜𝑏𝑗2X_{obj}^{2}italic_X start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, \cdots, XobjNsuperscriptsubscript𝑋𝑜𝑏𝑗𝑁X_{obj}^{N}italic_X start_POSTSUBSCRIPT italic_o italic_b italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT}, following the order of visual tags in the images.However, we find that simply prompting the model to list items in a zero-shot manner could lead to noisy and biased generation results, where the model may refer the tag to a distant object that is easy to describe. (see examples inSectionA.4).To mitigate this problem, we seek two complementary solutions: (1) We modify the system message of GPT-4V to avoid assigning tags to distant objects. (2) We manually design a few correct listing samples via human annotations, and use them as seed examples for in-context-learning to query GPT-4V.The details of our template is in Appendix.

In addition to listing, we also consider conversational data similar to LLaVA(Liu etal., 2023b), where GPT-4V is asked to generate mulit-turn question answering between an AI assistant and a person asking questions about the photo.Given a tagged image ITsuperscript𝐼𝑇I^{T}italic_I start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT, we use GPT-4V to generate instruction-following data in the form of {Person:ITsuperscript𝐼TI^{\texttt{T}}italic_I start_POSTSUPERSCRIPT T end_POSTSUPERSCRIPT XQ(i)superscriptsubscript𝑋Q𝑖X_{\texttt{Q}}^{(i)}italic_X start_POSTSUBSCRIPT Q end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT, Assistant: XA(i)superscriptsubscript𝑋A𝑖X_{\texttt{A}}^{(i)}italic_X start_POSTSUBSCRIPT A end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT}.

4.4 Model Training

We take the pretrained stage of LLaVA-1.5(Liu etal., 2023a) as the base model, and continue finetuning by mixing instruction tuning data of LLaVA-1.5 with our collected visual prompting data.For SoM-listing, we create 40 task templates as human instructions (e.g., “please enumerate object names in the tagged image”), and treat them as standard conversational data.We use the same training objective of next-token prediction to train general QA, SoM-QA and SoM-listing data.Specifically, we maximize the conditional log likelihood as follows:

logp(XA|Xv,XQ)=logi=1LpΘ(xi|I/IT,XQ,<i,XA,<i),𝑝conditionalsubscript𝑋Asubscript𝑋vsubscript𝑋Qsuperscriptsubscriptproduct𝑖1𝐿subscript𝑝Θconditionalsubscript𝑥𝑖𝐼superscript𝐼Tsubscript𝑋Qabsent𝑖subscript𝑋Aabsent𝑖-\log p(X_{\texttt{A}}|X_{\texttt{v}},X_{\texttt{Q}})=-\log\prod_{i=1}^{L}p_{%\Theta}(x_{i}|I/I^{\texttt{T}},X_{\texttt{Q},<i},X_{\texttt{A},<i}),- roman_log italic_p ( italic_X start_POSTSUBSCRIPT A end_POSTSUBSCRIPT | italic_X start_POSTSUBSCRIPT v end_POSTSUBSCRIPT , italic_X start_POSTSUBSCRIPT Q end_POSTSUBSCRIPT ) = - roman_log ∏ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_L end_POSTSUPERSCRIPT italic_p start_POSTSUBSCRIPT roman_Θ end_POSTSUBSCRIPT ( italic_x start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT | italic_I / italic_I start_POSTSUPERSCRIPT T end_POSTSUPERSCRIPT , italic_X start_POSTSUBSCRIPT Q , < italic_i end_POSTSUBSCRIPT , italic_X start_POSTSUBSCRIPT A , < italic_i end_POSTSUBSCRIPT ) ,(1)

where ΘΘ\Thetaroman_Θ are the trainable model parameters, XQ,<isubscript𝑋Qabsent𝑖X_{\texttt{Q},<i}italic_X start_POSTSUBSCRIPT Q , < italic_i end_POSTSUBSCRIPT and XA,<isubscript𝑋Aabsent𝑖X_{\texttt{A},<i}italic_X start_POSTSUBSCRIPT A , < italic_i end_POSTSUBSCRIPT are the instruction and answer tokens in all previous turns of conversations before the current prediction token xisubscript𝑥𝑖x_{i}italic_x start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT. The input image is I𝐼Iitalic_I or ITsuperscript𝐼TI^{\texttt{T}}italic_I start_POSTSUPERSCRIPT T end_POSTSUPERSCRIPT for LLaVA or SoM data, respectively.

A New Data Source and Learning Paradigm for Multimodal LLMs (3)
A New Data Source and Learning Paradigm for Multimodal LLMs (4)

5 Experiments

5.1 Experimental Settings

Experiment overview. We validate the method effectiveness from two aspects. First, inSection5.2, we benchmark the model’s capabilities in understand and describing SoM visual prompting. We design the tag listing task on MS-COCO to test the SoM performance. Second, inSection5.3, we evaluate if our dataset and model can benefit visual reasoning tasks, where we consider five representative visual question answering and reasoning tasks detailed as follows.

MLLM benchmarks. We consider the following multimodal LLM benchmarks in Table2 to validate SoM visual prompting’s benefit on visual reasoning.POPE(Li etal., 2023e) is carefully designed to evaluate object hallucination in multimodal LLMs. We follow POPE and report the F1 Score for the binary choice questions.MME(Fu etal., 2023) contains 2800 binary choice questions for perception and cognition evaluation. We report the overall perception score for the evaluated models.SEED-Bench(Li etal., 2023a) contains 19K multiple choice questions covering both image and video modality. We follow a previous study(Lin etal., 2023b) that reports the multiple choice accuracy on the image subset of 14k images, namely SEED-I.LLaVA-W: LLaVA-Bench (In-the-Wild)(Liu etal., 2023b) and MM-Vet(Yu etal., 2023) computes the evaluation score by prompting a GPT-4 based evaluator(OpenAI, 2023b) with both the predicted and ground-truth reference answer. The score is then scaled to the range of 0 to 100.We introduce extra implementation details inSectionA.1.

5.2 Evaluation on Tag Listing

First, we evaluate model performance on the tag listing task, aiming to answer two research questions: (1) Do model sizes matter in terms of learning SoM ability? (2) How will different sets of extra training data impact the SoM performance?We design the listing data based on images with ground-truth mask annotations from MS-COCO, and enumerate each object with corresponding class name. An example list is “1. person, 2. cat, 3. dog.”. We compute list-wise accuracy, where for a caption with N𝑁Nitalic_N items, the score is MN𝑀𝑁\frac{M}{N}divide start_ARG italic_M end_ARG start_ARG italic_N end_ARG with M𝑀Mitalic_M items predicted correctly by the model. With human annotation of objects in an image, we can automatically create abundant rule-based data (up to 100k) for studying model behaviors and perform quantitative evaluations.

For the first question, we find that larger LLM performs better for the listing task (seeFigure3(a)), presumably benefiting from the stronger language prior to help learn SoM prompting.For the second question, we decompose the 665k instruction data from LLaVA-1.5(Liu etal., 2023a) into two parts. We find that both general caption-QA data, as well as OCR-text data contribute to learning SoM ability when limited listing data are available (10k). The reason could be that OCR can help with identifying numeric tags, and general caption may help the model to recognize objects within an image, both of them are fundamental abilities required by SoM. In general, other visual instruction data may benefit learning SoM, especially when SoM data is scarce.

Overall, we observe that with only 10k data, we can outperform zero-shot GPT-4V in listing accuracy, whereas growing data size from 50k to 100k only slightly improves the listing performance. These findings suggest that collecting a small amount of data may be sufficient for learning SoM prompting.

MethodLLMRes.Pre-DataIT-DataPOPEMMESEED-ILLaVA-WMM-Vet
BLIP-2Vicuna-13B224129M-85.31293.849.738.122.4
InstructBLIPVicuna-7B224129M1.2M58.860.926.2
InstructBLIPVicuna-13B224129M1.2M78.91212.858.225.6
Fuyu-8BFuyu-8B60074.1728.621.4
LLaMA-Adapter-V2LLaMA2-7B3361328.435.2
mPLUG-Owl-2LLaMA2-7B448348M1450.264.136.2
Qwen-VLQwen-7B4481.4B50M62.3
Qwen-VL-ChatQwen-7B4481.4B50M1487.565.4
SPHINXLLaMA2-7B224--80.71476.169.173.536.0
LLaVA-1.5Vicuna-7B336558K665K85.91510.764.863.430.5
LLaVA-1.5Vicuna-13B336558K665K85.91531.368.270.735.4
SoM-LLaVA-1.5Vicuna-13B336558K695K86.61563.169.675.335.9
SoM-LLaVA-1.5-TVicuna-13B336558K695K87.01572.869.573.337.2

5.3 Evaluation on MLLM Benchmarks

We then train LLaVA-1.5 on our collected dataset and perform evaluation on MLLM benchmarks.As shown inTable2, we observe that our SoM-LLaVA-1.5, which is trained with a mixture of LLaVA visual instructions and our SoM data in order to learn SoM prompting, also obtains superior performance on general MLLM tasks.Surprisingly, we find that even without tagged images, SoM-LLaVA still attains strong performance and substantial improvement over the orignal LLaVA. This indicates the quality of our data and the potential of introducing listing data into general MLLM training to improve visual understanding and reasoning, as well as reduce hallucinations.We conjecture the reason that the great performance of SoM-LLaVA on non-tagged images is that “listing items one by one” with visual prompting guides the model to learn fine-grained semantics for image features. Related case studies and visualizations are inFigure8.For the performance of open-vocabulary listing, we present examples inSectionA.3.

5.4 Ablation Study on Mixture of Datasets

Finally, we perform ablation on different data mixture strategies inTable3. We consider mixing our listing and QA data generated fromSection4.3 with LLaVA-665k(Liu etal., 2023a), trained separately or together.Empirically, we find that mixing listing and QA data yields the best overall performance.InSection5.2, we find OCR data can help the learning of listing. Here we also notice that “listing item one by one” can in turn greatly improve the performance of OCR related task.The results on POPE indicates our data leads to lower hallucinations compared with ShareGPT-4V, which is a dense caption dataset without visual prompting.Placing tags on the images can seamlessly encode spatial information into the data for MLLMs to learn fine-grained vision language alignment.

Data CompositionData SizePOPEMMESEED-I
randompopularadversarialOCRoveralloverall
LLaVA-IT665K87.186.284.5125.01531.368.2
LLaVA-IT + Listing665K + 10k87.386.384.8147.51588.268.9
LLaVA-IT + QA695K + 20k87.586.484.7110.01540.069.2
LLaVA-IT + Listing + QA695K + 30k87.886.785.2140.01563.169.6
LLaVA-IT + ShareGPT-4V695K + 20k87.186.084.3110.01528.769.3

6 Analysis

A New Data Source and Learning Paradigm for Multimodal LLMs (5)

6.1 Probing Trained Models

We first analyze the tag-listing capacity of SoM-LLaVA-1.5 acquired through fine-tuning.In Figure 4, we show the attention maps on the five tagged objects, which are extracted from SoM-LLaVA-1.5 and LLaVA-1.5 respectively.The comparative example showcases that although both models can locate their model attention on the mentioned objects to some extent,the fine-tuned SoM-LLaVA-1.5 model can attend to and focus on characteristic regions of the object, which can also be accurately guided by the numeric ID tags.For example, the comparative attention maps on the object “Laptop” tagged with number 1 show that SoM-LLaVA-1.5 can clearly attend to the mentioned object with its main focus.In contrast, LLaVA-1.5 mistakenly attends to the monitor instead of the laptop, due to high similarity between these two objects.

In addition, we also observe that SoM-LLaVA-1.5 can be efficiently guided by the numeric ID tags to focus on the specific object the user refers to, even with multiple similar objects within the image.For example, the attention map of SoM-LLaVA-1.5 on the “Chair” tagged with a number 2 is mostly focusing on the chair on the left-hand side, instead of the similar chair on the right-hand side.SoM prompting in SoM-LLaVA-1.5 with such the capacity to accurately locate the tagged object, enables more flexible and easier user-referring queries without complicated language descriptions. The attention maps also verify our early hypothesis regarding the implicit association among thetext, tag, and object in SoM prompting.

A New Data Source and Learning Paradigm for Multimodal LLMs (6)
A New Data Source and Learning Paradigm for Multimodal LLMs (7)

6.2 Visual Reasoning with SoM Prompting

We present two examples of different models reasoning over the tagged images.In Figure 6, we examine a multi-step visual reasoning question (i.e., “Whose pants’ color is the same as someone else’s white shirt”),which requires the MLLM to first identify the mentioned objects (i.e., pants and shirt) and compare their visual features (i.e., the same white color).We observe from Figure 6 that LLaVA-1.5 provides an incorrect answer by falsely recognizing the person who wears the white shirt as a female.Such an incorrect answer can be caused by the inferior object recognition capacity in LLaVA-1.5.Similar observation from GPT-4V in Figure 6 showcases that incorrect recognition of the white color of the male’s pants can also cause incorrect reasoning conclusions in GPT-4V.In contrast, SoM-LLaVA-1.5 successfully identifies tags 1 and 9 with the same color in those image regions, while recognizing the two objects as white pants and white shirt, respectively. We show another example of tag selection in Figure6.

7 Conclusion

In this paper, we study SoM prompting of multimodal LLMs. We collected a tailored dataset that helps MLLMs acquiring the SoM visual prompting ability. Our approach demonstrates that MLLMs can learn SoM prompting using a small set of GPT-4V generated data, where the text describes the visual objects following the order of tags in the image. We then verify the effectiveness of SoM prompting on general VL reasoning tasks. Our enhanced model, SoM-LLaVA, consistently outperforms the original LLaVA model across five MLLM benchmarks.Our dataset and models will be released to facilitate vision and language research.

References

  • Alayrac etal. (2022)Jean-Baptiste Alayrac, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Yana Hasson, Karel Lenc, Arthur Mensch, Katherine Millican, Malcolm Reynolds, etal.Flamingo: a visual language model for few-shot learning.Advances in Neural Information Processing Systems, 35:23716–23736, 2022.
  • Bai etal. (2023)Jinze Bai, Shuai Bai, Shusheng Yang, Shijie Wang, Sinan Tan, Peng Wang, Junyang Lin, Chang Zhou, and Jingren Zhou.Qwen-vl: A frontier large vision-language model with versatile abilities.arXiv preprint arXiv:2308.12966, 2023.
  • Chen etal. (2023)Lin Chen, Jisong Li, Xiaoyi Dong, Pan Zhang, Conghui He, Jiaqi Wang, Feng Zhao, and Dahua Lin.Sharegpt4v: Improving large multi-modal models with better captions.arXiv preprint arXiv:2311.12793, 2023.
  • Chiang etal. (2023)Wei-Lin Chiang, Zhuohan Li, ZiLin, Ying Sheng, Zhanghao Wu, Hao Zhang, Lianmin Zheng, Siyuan Zhuang, Yonghao Zhuang, JosephE Gonzalez, etal.Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality.See https://vicuna. lmsys. org (accessed 14 April 2023), 2(3):6, 2023.
  • Dai etal. (2023)Wenliang Dai, Junnan Li, Dongxu Li, Anthony MengHuat Tiong, Junqi Zhao, Weisheng Wang, Boyang Li, Pascale Fung, and Steven Hoi.Instructblip: Towards general-purpose vision-language models with instruction tuning.arXiv preprint arXiv:2305.06500, 2023.
  • Deng etal. (2009)Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and LiFei-Fei.Imagenet: A large-scale hierarchical image database.In CVPR, 2009.
  • Dosovitskiy etal. (2020)Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby.An image is worth 16x16 words: Transformers for image recognition at scale.ArXiv, abs/2010.11929, 2020.URL https://api.semanticscholar.org/CorpusID:225039882.
  • Fu etal. (2023)Chaoyou Fu, Peixian Chen, Yunhang Shen, Yulei Qin, Mengdan Zhang, XuLin, Zhenyu Qiu, Wei Lin, Jinrui Yang, Xiawu Zheng, etal.Mme: A comprehensive evaluation benchmark for multimodal large language models.arXiv preprint arXiv:2306.13394, 2023.
  • Gao etal. (2023)Peng Gao, Jiaming Han, Renrui Zhang, Ziyi Lin, Shijie Geng, Aojun Zhou, Wei Zhang, Pan Lu, Conghui He, Xiangyu Yue, etal.Llama-adapter v2: Parameter-efficient visual instruction model.arXiv preprint arXiv:2304.15010, 2023.
  • Kazemzadeh etal. (2014)Sahar Kazemzadeh, Vicente Ordonez, Mark Matten, and Tamara Berg.Referitgame: Referring to objects in photographs of natural scenes.In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp. 787–798, 2014.
  • Kirillov etal. (2023)Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, AlexanderC Berg, Wan-Yen Lo, etal.Segment anything.arXiv preprint arXiv:2304.02643, 2023.
  • Li etal. (2023a)Bohao Li, Rui Wang, Guangzhi Wang, Yuying Ge, Yixiao Ge, and Ying Shan.Seed-bench: Benchmarking multimodal llms with generative comprehension.arXiv preprint arXiv:2307.16125, 2023a.
  • Li etal. (2023b)Chunyuan Li, Zhe Gan, Zhengyuan Yang, Jianwei Yang, Linjie Li, Lijuan Wang, and Jianfeng Gao.Multimodal foundation models: From specialists to general-purpose assistants.arXiv preprint arXiv:2309.10020, 2023b.
  • Li etal. (2023c)Feng Li, Hao Zhang, Peize Sun, Xueyan Zou, Shilong Liu, Jianwei Yang, Chunyuan Li, Lei Zhang, and Jianfeng Gao.Semantic-sam: Segment and recognize anything at any granularity.arXiv preprint arXiv:2307.04767, 2023c.
  • Li etal. (2023d)Junnan Li, Dongxu Li, Silvio Savarese, and Steven Hoi.Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models.arXiv preprint arXiv:2301.12597, 2023d.
  • Li etal. (2020)Xiujun Li, XiYin, Chunyuan Li, Xiaowei Hu, Pengchuan Zhang, Lei Zhang, Lijuan Wang, Houdong Hu, LiDong, Furu Wei, etal.Oscar: Object-semantics aligned pre-training for vision-language tasks.In ECCV, 2020.
  • Li etal. (2023e)Yifan Li, Yifan Du, Kun Zhou, Jinpeng Wang, WayneXin Zhao, and Ji-Rong Wen.Evaluating object hallucination in large vision-language models.arXiv preprint arXiv:2305.10355, 2023e.
  • Lin etal. (2023a)Kevin Lin, Faisal Ahmed, Linjie Li, Chung-Ching Lin, Ehsan Azarnasab, Zhengyuan Yang, Jianfeng Wang, Lin Liang, Zicheng Liu, Yumao Lu, etal.Mm-vid: Advancing video understanding with gpt-4v (ision).arXiv preprint arXiv:2310.19773, 2023a.
  • Lin etal. (2014)Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and CLawrence Zitnick.Microsoft coco: Common objects in context.In ECCV, 2014.
  • Lin etal. (2023b)Ziyi Lin, Chris Liu, Renrui Zhang, Peng Gao, Longtian Qiu, Han Xiao, Han Qiu, Chen Lin, Wenqi Shao, Keqin Chen, etal.Sphinx: The joint mixing of weights, tasks, and visual embeddings for multi-modal large language models.arXiv preprint arXiv:2311.07575, 2023b.
  • Liu etal. (2023a)Haotian Liu, Chunyuan Li, Yuheng Li, and YongJae Lee.Improved baselines with visual instruction tuning, 2023a.
  • Liu etal. (2023b)Haotian Liu, Chunyuan Li, Qingyang Wu, and YongJae Lee.Visual instruction tuning.In NeurIPS, 2023b.
  • Liu etal. (2024)Haotian Liu, Chunyuan Li, Yuheng Li, BoLi, Yuanhan Zhang, Sheng Shen, and YongJae Lee.Llava-next: Improved reasoning, ocr, and world knowledge, January 2024.URL https://llava-vl.github.io/blog/2024-01-30-llava-next/.
  • Mani etal. (2020)Arjun Mani, Nobline Yoo, Will Hinthorn, and Olga Russakovsky.Point and ask: Incorporating pointing into visual question answering.arXiv preprint arXiv:2011.13681, 2020.
  • Mao etal. (2016)Junhua Mao, Jonathan Huang, Alexander Toshev, Oana Camburu, AlanL Yuille, and Kevin Murphy.Generation and comprehension of unambiguous object descriptions.In CVPR, 2016.
  • OpenAI (2023a)OpenAI.Gpt-4v(ision) system card.2023a.URL https://cdn.openai.com/papers/GPTV_System_Card.pdf.
  • OpenAI (2023b)OpenAI.Gpt-4 technical report, 2023b.
  • Peng etal. (2023)Zhiliang Peng, Wenhui Wang, LiDong, Yaru Hao, Shaohan Huang, Shuming Ma, and Furu Wei.Kosmos-2: Grounding multimodal large language models to the world.arXiv preprint arXiv:2306.14824, 2023.
  • Radford etal. (2021)Alec Radford, JongWook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, etal.Learning transferable visual models from natural language supervision.arXiv preprint arXiv:2103.00020, 2021.
  • Rasheed etal. (2023)Hanoona Rasheed, Muhammad Maaz, Sahal Shaji, Abdelrahman Shaker, Salman Khan, Hisham Cholakkal, RaoM Anwer, Erix Xing, Ming-Hsuan Yang, and FahadS Khan.Glamm: Pixel grounding large multimodal model.arXiv preprint arXiv:2311.03356, 2023.
  • Schuhmann etal. (2022)Christoph Schuhmann, Romain Beaumont, CadeW Gordon, Ross Wightman, Theo Coombes, Aarush Katta, Clayton Mullis, Patrick Schramowski, SrivatsaR Kundurthy, Katherine Crowson, etal.Laion-5b: An open large-scale dataset for training next generation image-text models.In Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2022.
  • Shtedritski etal. (2023)Aleksandar Shtedritski, Christian Rupprecht, and Andrea Vedaldi.What does clip know about a red circle? visual prompt engineering for vlms.arXiv preprint arXiv:2304.06712, 2023.
  • Team etal. (2023)Gemini Team, Rohan Anil, Sebastian Borgeaud, Yonghui Wu, Jean-Baptiste Alayrac, Jiahui Yu, Radu Soricut, Johan Schalkwyk, AndrewM Dai, Anja Hauth, etal.Gemini: a family of highly capable multimodal models.arXiv preprint arXiv:2312.11805, 2023.
  • Touvron etal. (2023)Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, etal.Llama 2: Open foundation and fine-tuned chat models.arXiv preprint arXiv:2307.09288, 2023.
  • Wang etal. (2023a)Weihan Wang, Qingsong Lv, Wenmeng Yu, Wenyi Hong, JiQi, Yan Wang, Junhui Ji, Zhuoyi Yang, Lei Zhao, Xixuan Song, etal.Cogvlm: Visual expert for pretrained language models.arXiv preprint arXiv:2311.03079, 2023a.
  • Wang etal. (2023b)Wenhai Wang, Zhe Chen, Xiaokang Chen, Jiannan Wu, Xizhou Zhu, Gang Zeng, Ping Luo, Tong Lu, Jie Zhou, YuQiao, etal.Visionllm: Large language model is also an open-ended decoder for vision-centric tasks.arXiv preprint arXiv:2305.11175, 2023b.
  • Wei etal. (2022)Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, EdChi, Quoc Le, and Denny Zhou.Chain of thought prompting elicits reasoning in large language models.arXiv preprint arXiv:2201.11903, 2022.
  • Yan etal. (2023a)AnYan, Zhankui He, Jiacheng Li, Tianyang Zhang, and Julian McAuley.Personalized showcases: Generating multi-modal explanations for recommendations.In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 2251–2255, 2023a.
  • Yan etal. (2023b)AnYan, Zhengyuan Yang, Wanrong Zhu, Kevin Lin, Linjie Li, Jianfeng Wang, Jianwei Yang, Yiwu Zhong, Julian McAuley, Jianfeng Gao, etal.Gpt-4v in wonderland: Large multimodal models for zero-shot smartphone gui navigation.arXiv preprint arXiv:2311.07562, 2023b.
  • Yang etal. (2023a)Jianwei Yang, Hao Zhang, Feng Li, Xueyan Zou, Chunyuan Li, and Jianfeng Gao.Set-of-mark prompting unleashes extraordinary visual grounding in gpt-4v.arXiv preprint arXiv:2310.11441, 2023a.
  • Yang etal. (2021)Zhengyuan Yang, Yijuan Lu, Jianfeng Wang, XiYin, Dinei Florencio, Lijuan Wang, Cha Zhang, Lei Zhang, and Jiebo Luo.Tap: Text-aware pre-training for text-vqa and text-caption.In CVPR, pp. 8751–8761, 2021.
  • Yang etal. (2023b)Zhengyuan Yang, Linjie Li, Kevin Lin, Jianfeng Wang, Chung-Ching Lin, Zicheng Liu, and Lijuan Wang.The dawn of lmms: Preliminary explorations with gpt-4v (ision).arXiv preprint arXiv:2309.17421, 2023b.
  • Ye etal. (2023)Qinghao Ye, Haiyang Xu, Guohai Xu, Jiabo Ye, Ming Yan, Yiyang Zhou, Junyang Wang, Anwen Hu, Pengcheng Shi, Yaya Shi, etal.mplug-owl: Modularization empowers large language models with multimodality.arXiv preprint arXiv:2304.14178, 2023.
  • Yu etal. (2016)Licheng Yu, Patrick Poirson, Shan Yang, AlexanderC Berg, and TamaraL Berg.Modeling context in referring expressions.In ECCV, 2016.
  • Yu etal. (2023)Weihao Yu, Zhengyuan Yang, Linjie Li, Jianfeng Wang, Kevin Lin, Zicheng Liu, Xinchao Wang, and Lijuan Wang.Mm-vet: Evaluating large multimodal models for integrated capabilities.arXiv preprint arXiv:2308.02490, 2023.
  • Zhang etal. (2023)Jiangning Zhang, Xuhai Chen, Zhucun Xue, Yabiao Wang, Chengjie Wang, and Yong Liu.Exploring grounding potential of vqa-oriented gpt-4v for zero-shot anomaly detection.arXiv preprint arXiv:2311.02612, 2023.
  • Zhu etal. (2022)Wanrong Zhu, AnYan, Yujie Lu, Wenda Xu, XinEric Wang, Miguel Eckstein, and WilliamYang Wang.Visualize before you write: Imagination-guided open-ended text generation.arXiv preprint arXiv:2210.03765, 2022.
  • Zou etal. (2023)Xueyan Zou, Jianwei Yang, Hao Zhang, Feng Li, Linjie Li, Jianfeng Gao, and YongJae Lee.Segment everything everywhere all at once.arXiv preprint arXiv:2304.06718, 2023.

Appendix A Appendix

A.1 Implementation details.

The LLaVA-1.5 model contains a CLIP-ViT-L-336px visual encoder(Radford etal., 2021) and a Vicuna-7/13B language model(Chiang etal., 2023), connected by an MLP projection layer.Our main experiments are conducted on 8X and 4X 80GB A100 GPUs for llava-13b and llava-7b models, with a batch size of 128 and 64, respectively.We collected 10k SoM-listing data and 20k SoM-QA data using GPT-4V turbo.For visual tagging, we use the level-2 granularity of Semantic SAM to annotate all images from MS-COCO, to learn fine-grained object-text alignment. During inference, we find that the existing MLLM benchmarks mostly consist of high-level questions about an image, and level-1 annotation with fewer tags works better.

We report results of following MLLMs on public benchmarks: BLIP-2(Li etal., 2023d), InstructBLIP(Dai etal., 2023), Fuyu-8B111https://www.adept.ai/blog/fuyu-8b, LLaMA-Adapter-V2(Gao etal., 2023), mPLUG-Owl-2(Ye etal., 2023), Qwen-VL(Bai etal., 2023), SPHINX(Lin etal., 2023b), and LLaVA-1.5(Liu etal., 2023a).

A.2 Comparison Results on Reasoning on Images without Tags

We additionally analyze how LLaVA-1.5 and SoM-LLaVA-1.5 perform differently when images with no tags are provided.In Figure 8 and Figure 8 we can observe that the discrepancies between the attention maps extracted from the two models in both cases are relatively insignificant. Such observation suggests that LLaVA-1.5 has pre-trained with good multimodal cross-attention that enables the MLLM to capture the most characteristic visual features in the images.However, due to the lack of alignment between visual semantics and textual semantics, MLLMs like LLaVA-1.5 may not correctly associate textual information with relevant visual evidence, which further causes incorrect answers in visual reasoning.With SoM fine-tuning, we reinforce the MLLM’s visual understanding of specific objects in the image by asking the model to list objects one by one.By bridging the objects’ visual features and their semantic meanings, the MLLM can better refer to the visual objects and answer the questions with more accurate object descriptions.

A New Data Source and Learning Paradigm for Multimodal LLMs (8)
A New Data Source and Learning Paradigm for Multimodal LLMs (9)

A.3 Listing Comparison from SoM-LLaVa and GPT-4V.

We present examples of open vocabulary listing with our SoM-LLaVA and GPT-4V, as shown inFigure10 andFigure10.

A New Data Source and Learning Paradigm for Multimodal LLMs (10)
A New Data Source and Learning Paradigm for Multimodal LLMs (11)

A.4 GPT-4V listings with Different Prompting Methods

We present the listing results from GPT-4V with different prompting methods, as shown inTable4 andTable5. 2-shot in-context learning leads to more accurate listings.

Listing example from GPT-4V, woman by the water.
A New Data Source and Learning Paradigm for Multimodal LLMs (12)
GPT-4 zero-shot- Tag 1 is on the water of a lake or river.- Tag 2 is on a boat.- Tag 3 is on a swimming platform or dock in the water.- Tag 4 is on the land or shore.- Tag 5 is on an informational sign or a notice board.- Tag 6 is on a tree.- Tag 7 is on a pink parasol or umbrella.- Tag 8 is on a bathing suit.- Tag 9 is placed where the blurred face is, describing no item.
GPT-4V improved sys-msgHere are the items tagged in the image:1. Blue water, possibly a lake or a river.2. The lower part of a white, vertical pole, likely a part of the structure tagged with 5.3. Green foliage or trees in the distance across the water.4. The edge of a large white sign on a poll, with some text visible but not readable.5. A white lifeguard stand or chair with blue accents and a sign attached.6. Blue sky.7. Blue sky with clouds.8. The midsection of a person wearing a floral swimsuit.9. A pink umbrella, held by the person whose face is blurred.The face of the person holding the umbrella is blurred for privacy reasons.
GPT-4V 2-shot learningHere’s a list of the tagged items in the image:1. The bottom edge of a body of water, likely a lake or river given the surrounding context.2. Lower arm and hand of a person holding an umbrella. The person’s hand is specifically at the center of this tag.3. Green foliage in the distance, on the other side of the water, which could be trees on the far shore.4. The tag appears to be on the lower part of the torso of a person, specifically around the waist area.5. A white structure with blue markings, which appears to be part of a lifeguard stand or beach-related signage.6. A sign, partially obscured by the tag, with written text.7. Clear blue sky.8. The tag is centered on the midsection of a person wearing a floral swimsuit.9. A pink parasol or umbrella, which the person is holding.

Listing example from GPT-4V, flower and vase.
A New Data Source and Learning Paradigm for Multimodal LLMs (13)
GPT-4 zero-shot1. White vase2. Small white flowers (species not clearly identifiable due to focus)3. Green foliage or plant (specific species unclear)4. Green plant (specific species unclear)5. Plant stems (specific species unclear)6. Green leaves (specific species unclear)7. Green leaves (specific species unclear)8. Green foliage or plant (specific species unclear)9. Green foliage or plant in the background (specific species unclear due to focus)
GPT-4V improved sys-msg1. A white vase.2. Part of a shrub or bush, possibly with tiny white flowers, in the background.3. Greenery that looks like a lawn or ground cover, it’s blurred and behind a ledge or step with tag number 8.4. Some plants which are blurred, they could be grass or small plants.5. Flowers, specifically white and possibly some type of hydrangea.6. Green foliage of a tree or large bush in the background.7. The pedestal or foot of the vase from tag number 1.8. This tag seems to be positioned on the edge of a railing or a surface with a white finish.9. More blurred greenery, possibly bushes or trees in the background.
GPT-4V 2-shot learning1. A white vase with a textured design.2. Tag 2 is hovering above some sort of vegetation, likely a bush or shrub.3. The tag is pointing to an area that appears to be part of a well-manicured lawn or grassy area.4. Green grass or a plant is suggested by this tag, although it is not entirely clear.5. This appears to be pointing to flowers in the bouquet, possibly white Hydrangeas.6. Appears to be pointing to a shadow cast on the lawn or a shaded area of grass.7. A cluster of flowers within the vase; they are likely white hydrangeas or another type of similar-looking flower.8. Is tagging the flat surface which appears to be a ledge or balustrade, likely made out of painted wood based on the context.9. A blurred area, suggestive of more green foliage or shrubbery in the background.

A.5 SoM Granularity Analysis

We present examples of visual tagging with semantic-SAM anddifferent granularity, as shown inFigures14, 14, 14 and14.

A New Data Source and Learning Paradigm for Multimodal LLMs (14)
A New Data Source and Learning Paradigm for Multimodal LLMs (15)
A New Data Source and Learning Paradigm for Multimodal LLMs (16)
A New Data Source and Learning Paradigm for Multimodal LLMs (17)

A.6 SoM Data in Existing Training Sources

A New Data Source and Learning Paradigm for Multimodal LLMs (18)

Tables6, 7 and8 shows a few examples that consist of listing in the text content.

ImageText
A New Data Source and Learning Paradigm for Multimodal LLMs (19)Question:What are some potential benefits or reasons people might gather at the beach to fly kites or engage in kiteboarding?
Answer:There are several reasons or benefits for people to gather at a beach to fly kites or engage in kiteboarding:1. Socializing and bonding: Kite flying and kiteboarding events can bring people together, fostering a sense of community and social interaction among the participants. These activities allow people of all ages and backgrounds to establish connections and develop new friendships.2. Outdoor recreation: Engaging in kite flying or kiteboarding can be a fun way to spend time outdoors and enjoy the fresh air and natural scenery. Beaches often offer wide-open spaces and consistent winds, which are ideal conditions for these activities.3. Physical exercise: Both kite flying and kiteboarding can provide a form of physical exercise, improving the participants’ overall fitness and well-being. Kite flying can be a low-intensity workout, strengthening the arms and shoulders, while kiteboarding is a more physically demanding sport that can enhance balance, core strength, and cardiovascular endurance.4. Stress relief: Participating in kite flying or kiteboarding can help reduce stress and promote relaxation. These activities can serve as a form of mindfulness and mental exercise, allowing the participants to focus solely on the task at hand, temporarily setting aside everyday concerns and responsibilities.5. Skill development: Flying kites and kiteboarding involve the development and refinement of various skills, such as coordination, dexterity, and control. Mastering these skills can provide a sense of accomplishment and personal growth.In summary, people may gather at a beach to fly kites or engage in kiteboarding for various reasons, including socializing, outdoor recreation, exercise, stress relief, and skill development. These activities can offer a range of physical, mental, and social benefits, making them popular and enjoyable pastimes for those who take part.
ImageText
A New Data Source and Learning Paradigm for Multimodal LLMs (20)Question:Can you provide some background information about coal-powered locomotives?
Answer:Coal-powered locomotives, also known as steam locomotives, were the primary mode of railway transportation from the early 19th century until the mid-20th century. These locomotives used steam power generated by burning coal to operate their engines, which provided the force necessary for the train to move along the tracks.A coal-powered locomotive was usually comprised of several key components:1. A firebox or combustion chamber where coal was burned to create heat.2. A boiler that contained water, heated by the firebox, to produce steam.3. Pistons and cylinders that converted the steam pressure into mechanical motion, allowing the locomotive’s wheels to turn.4. A coal car or tender, which carried the coal fuel for the locomotive, usually located just behind the engine.During the first half of the 20th century, coal-powered locomotives were steadily replaced by diesel and electric trains, which proved to be much more efficient and cleaner. Although steam locomotives are largely obsolete today, many are preserved in museums or operated for tourist attractions to provide a glimpse into the history of railway transportation.The image of an old coal-powered locomotive and coal car on old tracks in a city setting offers a snapshot of a bygone era, showcasing the marvels of early railway technology and providing us with a window into the past.
ImageText
A New Data Source and Learning Paradigm for Multimodal LLMs (21)Question:Can you talk about the role of footwork, movement, and jumping in tennis?
Answer:Footwork, movement, and jumping are crucial components of a good tennis game, both in terms of enhancing overall performance and reducing the risk of injuries.1. Footwork:Good footwork allows a tennis player to move efficiently and quickly around the court, enabling them to reach the ball and return it effectively. Proper footwork involves maintaining a balanced stance, taking small adjustment steps, and employing various movement patterns such as the shuffle, crossover, and side-step. Mastering footwork not only improves speed and agility but also ensures the player is in the right position to execute their strokes with optimal power and accuracy.2. Movement:Movement is an essential aspect of tennis, as players must continually readjust their position to respond to their opponent’s shots. Efficient movement on the court requires a combination of speed, agility, flexibility, and endurance. Anticipation plays a significant role in effective movement, as players must predict the trajectory of the ball and adjust their position accordingly. Top players often possess the ability to change direction quickly and cover the court effectively, enabling them to counter their opponent’s strategies and make challenging shots.3. Jumping:Jumping is a valuable skill in tennis, particularly when executing specific shots such as the overhead smash, serve or aggressive baseline shots. A well-executed jump allows the player to generate more power, reach for higher balls, and create sharper angles in their shots. Jumping while hitting the ball gives the player a slight height advantage, which enables them to strike the ball at a higher contact point, resulting in a more aggressive and effective shot. However, it is essential to practice proper techniques and ensure a safe landing to avoid potential injuries.In conclusion, footwork, movement, and jumping are vital components of tennis performance. Mastering these skills makes a player more effective and efficient on the court, enhancing their ability to respond to various situations and execute a wide range of shots. Proper training and practice in these areas can significantly elevate a player’s game and contribute to overall success in the sport.
A New Data Source and Learning Paradigm for Multimodal LLMs (2024)
Top Articles
Latest Posts
Article information

Author: Jerrold Considine

Last Updated:

Views: 5301

Rating: 4.8 / 5 (78 voted)

Reviews: 85% of readers found this page helpful

Author information

Name: Jerrold Considine

Birthday: 1993-11-03

Address: Suite 447 3463 Marybelle Circles, New Marlin, AL 20765

Phone: +5816749283868

Job: Sales Executive

Hobby: Air sports, Sand art, Electronics, LARPing, Baseball, Book restoration, Puzzles

Introduction: My name is Jerrold Considine, I am a combative, cheerful, encouraging, happy, enthusiastic, funny, kind person who loves writing and wants to share my knowledge and understanding with you.