Dataset Viewer
image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000002301.jpg
|
2301_evjvqa
|
[{"from": "human", "value": "what color is the shirt of the girl wearing glasses?\n<image>"}, {"from": "gpt", "value": "the girl wearing glasses wears a red shirt"}]
|
00000002301.jpg
|
2301_evjvqa
|
[{"from": "human", "value": "what fruit is put on the table next to the girl in red t-shirt?\n<image>"}, {"from": "gpt", "value": "the girl in red t-shirt is standing next to the lychees table"}]
|
00000002301.jpg
|
2301_evjvqa
|
[{"from": "human", "value": "what are put on the table next to the woman in red shirt?\n<image>"}, {"from": "gpt", "value": "bunches of lychee"}]
|
00000002302.jpg
|
2302_evjvqa
|
[{"from": "human", "value": "what is this woman looking at?\n<image>"}, {"from": "gpt", "value": "she is looking at the milk"}]
|
00000002302.jpg
|
2302_evjvqa
|
[{"from": "human", "value": "what is the woman in white shirt doing?\n<image>"}, {"from": "gpt", "value": "the woman in white shirt is holding and looking at the bottle of milk"}]
|
00000002302.jpg
|
2302_evjvqa
|
[{"from": "human", "value": "what is the girl about to buy at this stall?\n<image>"}, {"from": "gpt", "value": "bottles of milk"}]
|
00000002304.jpg
|
2304_evjvqa
|
[{"from": "human", "value": "what is the girl in pink shirt wearing on her shoulder?\n<image>"}, {"from": "gpt", "value": "she is wearing a black purse"}]
|
00000002304.jpg
|
2304_evjvqa
|
[{"from": "human", "value": "where are those two peoples at together?\n<image>"}, {"from": "gpt", "value": "they are together in a supermarket"}]
|
00000002304.jpg
|
2304_evjvqa
|
[{"from": "human", "value": "what are the girl and the boy come to the market for?\n<image>"}, {"from": "gpt", "value": "shopping"}]
|
00000002308.jpg
|
2308_evjvqa
|
[{"from": "human", "value": "what is the girl in blue jacket coming to this store for?\n<image>"}, {"from": "gpt", "value": "she is coming for christmas decorations"}]
|
00000002308.jpg
|
2308_evjvqa
|
[{"from": "human", "value": "who is coming with the girl in blue jacket?\n<image>"}, {"from": "gpt", "value": "the girl in pink jacket"}]
|
00000002308.jpg
|
2308_evjvqa
|
[{"from": "human", "value": "how does the store owner look like?\n<image>"}, {"from": "gpt", "value": "the woman wearing blue t-shirt at the checkout"}]
|
00000002308.jpg
|
2308_evjvqa
|
[{"from": "human", "value": "what does the woman in blue t-shirt sell?\n<image>"}, {"from": "gpt", "value": "christmas decorations"}]
|
00000002310.jpg
|
2310_evjvqa
|
[{"from": "human", "value": "what is she doing at the supermarket?\n<image>"}, {"from": "gpt", "value": "scanning the qr code"}]
|
00000002311.jpg
|
2311_evjvqa
|
[{"from": "human", "value": "what would we want buy if we come to this store?\n<image>"}, {"from": "gpt", "value": "we would buy a motorbike if we come to this store"}]
|
00000002314.jpg
|
2314_evjvqa
|
[{"from": "human", "value": "what color is the board of the store?\n<image>"}, {"from": "gpt", "value": "the board of the store is painted orange and white"}]
|
00000002318.jpg
|
2318_evjvqa
|
[{"from": "human", "value": "how was the ambulance's license plate look like?\n<image>"}, {"from": "gpt", "value": "the ambulance plate is rectangular, blue background and white text"}]
|
00000002319.jpg
|
2319_evjvqa
|
[{"from": "human", "value": "what products are sold on the right aisle?\n<image>"}, {"from": "gpt", "value": "fresh vegestables"}]
|
00000002319.jpg
|
2319_evjvqa
|
[{"from": "human", "value": "what are the customers doing here?\n<image>"}, {"from": "gpt", "value": "they are shopping in the supermarket"}]
|
00000002319.jpg
|
2319_evjvqa
|
[{"from": "human", "value": "what color is the mask of the girl in white t-shirt?\n<image>"}, {"from": "gpt", "value": "the girl in white t-shirt wears a white mask"}]
|
00000002328.jpg
|
2328_evjvqa
|
[{"from": "human", "value": "what color is the masks of those three people?\n<image>"}, {"from": "gpt", "value": "they are wearing white masks"}]
|
00000002328.jpg
|
2328_evjvqa
|
[{"from": "human", "value": "how was the woman's shirt look like?\n<image>"}, {"from": "gpt", "value": "the woman wears a white shirt with black polka dots"}]
|
00000002328.jpg
|
2328_evjvqa
|
[{"from": "human", "value": "who is standing in this store?\n<image>"}, {"from": "gpt", "value": "there are three people, a woman and two men"}]
|
00000002333.jpg
|
2333_evjvqa
|
[{"from": "human", "value": "who is making a speech at this event?\n<image>"}, {"from": "gpt", "value": "the man with white hair"}]
|
00000002333.jpg
|
2333_evjvqa
|
[{"from": "human", "value": "what is the man with white hair standing for?\n<image>"}, {"from": "gpt", "value": "he is making a speech at the event?"}]
|
00000002333.jpg
|
2333_evjvqa
|
[{"from": "human", "value": "where is the man with white hair standing for the speech?\n<image>"}, {"from": "gpt", "value": "he is standing on the rostrum"}]
|
00000002338.jpg
|
2338_evjvqa
|
[{"from": "human", "value": "how were the cashier costumes look like?\n<image>"}, {"from": "gpt", "value": "white shirt and white hat"}]
|
00000002338.jpg
|
2338_evjvqa
|
[{"from": "human", "value": "what is the man in black t-shirt doing at the checkout?\n<image>"}, {"from": "gpt", "value": "he is packaging the goods into the plastic bag"}]
|
00000002338.jpg
|
2338_evjvqa
|
[{"from": "human", "value": "what color is the shirt of the man packaging the goods into the plastic bag?\n<image>"}, {"from": "gpt", "value": "the man packaging the goods into the plastic bag wears black t-shirt"}]
|
00000002339.jpg
|
2339_evjvqa
|
[{"from": "human", "value": "what is this store mainly sold?\n<image>"}, {"from": "gpt", "value": "this store mainly sells different kinds of dress"}]
|
00000002340.jpg
|
2340_evjvqa
|
[{"from": "human", "value": "which kid is wearing the bag?\n<image>"}, {"from": "gpt", "value": "the boy wearing black shirt is wearing the bag"}]
|
00000002340.jpg
|
2340_evjvqa
|
[{"from": "human", "value": "what are the kids doing at the book aisle?\n<image>"}, {"from": "gpt", "value": "they are looking for books"}]
|
00000002345.jpg
|
2345_evjvqa
|
[{"from": "human", "value": "what are those two man doing?\n<image>"}, {"from": "gpt", "value": "they are fixing the board for the store"}]
|
00000002345.jpg
|
2345_evjvqa
|
[{"from": "human", "value": "where are the men fixing the board?\n<image>"}, {"from": "gpt", "value": "one standing behind the board and one is next to the board"}]
|
00000002345.jpg
|
2345_evjvqa
|
[{"from": "human", "value": "what helps the men to climb up the store?\n<image>"}, {"from": "gpt", "value": "ladders"}]
|
00000002349.jpg
|
2349_evjvqa
|
[{"from": "human", "value": "where is the man sitting at?\n<image>"}, {"from": "gpt", "value": "he is sitting on the parking lot for disabled people"}]
|
00000002349.jpg
|
2349_evjvqa
|
[{"from": "human", "value": "where is the black car parking?\n<image>"}, {"from": "gpt", "value": "the black car is parking at the disabled people area"}]
|
00000002349.jpg
|
2349_evjvqa
|
[{"from": "human", "value": "how was the man sitting look like?\n<image>"}, {"from": "gpt", "value": "white t-shirt, black jeans and wearing black sneakers"}]
|
00000002356.jpg
|
2356_evjvqa
|
[{"from": "human", "value": "which side of the man with white hair is the woman standing?\n<image>"}, {"from": "gpt", "value": "the right side"}]
|
00000002359.jpg
|
2359_evjvqa
|
[{"from": "human", "value": "what color is the balloon?\n<image>"}, {"from": "gpt", "value": "the balloon is red"}]
|
00000002359.jpg
|
2359_evjvqa
|
[{"from": "human", "value": "what is the woman wearing on her shoulder?\n<image>"}, {"from": "gpt", "value": "a white handbag"}]
|
00000002359.jpg
|
2359_evjvqa
|
[{"from": "human", "value": "where is the trolley ?\n<image>"}, {"from": "gpt", "value": "the trolley is next to the tomatoes aisle"}]
|
00000002363.jpg
|
2363_evjvqa
|
[{"from": "human", "value": "where is the watch?\n<image>"}, {"from": "gpt", "value": "the watch is on the left hand of the person who is pushing the trolley"}]
|
00000002363.jpg
|
2363_evjvqa
|
[{"from": "human", "value": "what fruit on the trolley is yellow?\n<image>"}, {"from": "gpt", "value": "the banana"}]
|
00000002363.jpg
|
2363_evjvqa
|
[{"from": "human", "value": "what are the hands pushing?\n<image>"}, {"from": "gpt", "value": "the trolley"}]
|
00000002364.jpg
|
2364_evjvqa
|
[{"from": "human", "value": "what are the people gathering on the shore for?\n<image>"}, {"from": "gpt", "value": "to watch the boat racing"}]
|
00000002364.jpg
|
2364_evjvqa
|
[{"from": "human", "value": "what are the people holding the paddle at the same time for?\n<image>"}, {"from": "gpt", "value": "to race with other boats"}]
|
00000002364.jpg
|
2364_evjvqa
|
[{"from": "human", "value": "which boat is leading the race?\n<image>"}, {"from": "gpt", "value": "the boat of blue shirt members"}]
|
00000002370.jpg
|
2370_evjvqa
|
[{"from": "human", "value": "what products are sold at this area?\n<image>"}, {"from": "gpt", "value": "different kinds of bread"}]
|
00000002370.jpg
|
2370_evjvqa
|
[{"from": "human", "value": "what product can we find at this stall?\n<image>"}, {"from": "gpt", "value": "different kinds of bread"}]
|
00000002370.jpg
|
2370_evjvqa
|
[{"from": "human", "value": "who is carrying the kid in pink shirt on the trolley?\n<image>"}, {"from": "gpt", "value": "the man wearing black t-shirt"}]
|
00000002370.jpg
|
2370_evjvqa
|
[{"from": "human", "value": "what is the woman wearing black dress with white dots about to buy?\n<image>"}, {"from": "gpt", "value": "the woman is about to buy bread"}]
|
00000002371.jpg
|
2371_evjvqa
|
[{"from": "human", "value": "what special occasion are those decorations used for?\n<image>"}, {"from": "gpt", "value": "tet lunar new year"}]
|
00000002371.jpg
|
2371_evjvqa
|
[{"from": "human", "value": "what is the girl wearing jacket on her head is holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding her purse"}]
|
00000002375.jpg
|
2375_evjvqa
|
[{"from": "human", "value": "how many kids are there on the swings?\n<image>"}, {"from": "gpt", "value": "there are two kids sitting on the swings"}]
|
00000002375.jpg
|
2375_evjvqa
|
[{"from": "human", "value": "where are those kids playing at?\n<image>"}, {"from": "gpt", "value": "in a playground of the park"}]
|
00000002375.jpg
|
2375_evjvqa
|
[{"from": "human", "value": "what is the kid wearing cap doing?\n<image>"}, {"from": "gpt", "value": "playing on the swing"}]
|
00000002380.jpg
|
2380_evjvqa
|
[{"from": "human", "value": "what is the girl wearing on her wrist?\n<image>"}, {"from": "gpt", "value": "there is a watch on her wrist"}]
|
00000002380.jpg
|
2380_evjvqa
|
[{"from": "human", "value": "what is the woman's hand holding?\n<image>"}, {"from": "gpt", "value": "she is holding the flowers"}]
|
00000002380.jpg
|
2380_evjvqa
|
[{"from": "human", "value": "who is standing close to the girl?\n<image>"}, {"from": "gpt", "value": "the girl is standing alone "}]
|
00000002383.jpg
|
2383_evjvqa
|
[{"from": "human", "value": "what is the woman doing with her camera?\n<image>"}, {"from": "gpt", "value": "she is taking pictures"}]
|
00000002383.jpg
|
2383_evjvqa
|
[{"from": "human", "value": "how was the costumes of the girl standing?\n<image>"}, {"from": "gpt", "value": "she wears white shirt, black skirt and hat"}]
|
00000002383.jpg
|
2383_evjvqa
|
[{"from": "human", "value": "what device is the girl holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a camera"}]
|
00000002384.jpg
|
2384_evjvqa
|
[{"from": "human", "value": "what store is the girl in red dress standing at?\n<image>"}, {"from": "gpt", "value": "she is standing in a bookstore"}]
|
00000002384.jpg
|
2384_evjvqa
|
[{"from": "human", "value": "what is the girl in red dress holding on her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a book"}]
|
00000002384.jpg
|
2384_evjvqa
|
[{"from": "human", "value": "what are those coming to this bookstore for?\n<image>"}, {"from": "gpt", "value": "to buy books"}]
|
00000002385.jpg
|
2385_evjvqa
|
[{"from": "human", "value": "where are the life jacket placed on the boat?\n<image>"}, {"from": "gpt", "value": "it is placed at the back of the boat, behind the person in pink shirt"}]
|
00000002385.jpg
|
2385_evjvqa
|
[{"from": "human", "value": "what is the woman holding the paddles for?\n<image>"}, {"from": "gpt", "value": "she is holding the paddle to ride the boat"}]
|
00000002385.jpg
|
2385_evjvqa
|
[{"from": "human", "value": "what color is the shirt of the woman wearing conical hat?\n<image>"}, {"from": "gpt", "value": "the woman wearing conical hat is wearing a pink shirt"}]
|
00000002385.jpg
|
2385_evjvqa
|
[{"from": "human", "value": "where can we find the conical hat?\n<image>"}, {"from": "gpt", "value": "the conical hat is worn by the woman in pink shirt"}]
|
00000002386.jpg
|
2386_evjvqa
|
[{"from": "human", "value": "what is the staff in yellow t-shirt wearing on his hat?\n<image>"}, {"from": "gpt", "value": "he is wearing a yellow cap"}]
|
00000002386.jpg
|
2386_evjvqa
|
[{"from": "human", "value": "what are placing on the the table in front of the staff in yellow shirt?\n<image>"}, {"from": "gpt", "value": "breads and food"}]
|
00000002386.jpg
|
2386_evjvqa
|
[{"from": "human", "value": "how was the costumes of the man standing between two baskets of bread?\n<image>"}, {"from": "gpt", "value": "the costumes includes yellow t-shirt, black pants and yellow cap"}]
|
00000002389.jpg
|
2389_evjvqa
|
[{"from": "human", "value": "what is hung on the strings above?\n<image>"}, {"from": "gpt", "value": "the conical hats"}]
|
00000002389.jpg
|
2389_evjvqa
|
[{"from": "human", "value": "where can we find the voltage transformer?\n<image>"}, {"from": "gpt", "value": "the voltage transformer is on the left sidewalk of the road"}]
|
00000002389.jpg
|
2389_evjvqa
|
[{"from": "human", "value": "what is this girl doing in the middle of the road?\n<image>"}, {"from": "gpt", "value": "she is walking"}]
|
00000002390.jpg
|
2390_evjvqa
|
[{"from": "human", "value": "where is the bench in this room?\n<image>"}, {"from": "gpt", "value": "the bench is in the middle of the room"}]
|
00000002390.jpg
|
2390_evjvqa
|
[{"from": "human", "value": "how many paintinings are hanging on the walls?\n<image>"}, {"from": "gpt", "value": "there are six paintings hanging on the walls"}]
|
00000002390.jpg
|
2390_evjvqa
|
[{"from": "human", "value": "how many paintings are there on the black wall?\n<image>"}, {"from": "gpt", "value": "there is one painting on the black wall"}]
|
00000002392.jpg
|
2392_evjvqa
|
[{"from": "human", "value": "what can this aisle sell?\n<image>"}, {"from": "gpt", "value": "this aisle sells woman costumes"}]
|
00000002392.jpg
|
2392_evjvqa
|
[{"from": "human", "value": "where can we find the blue plastic chair?\n<image>"}, {"from": "gpt", "value": "behind the woman looking into the stalls"}]
|
00000002392.jpg
|
2392_evjvqa
|
[{"from": "human", "value": "what can we buy at this place?\n<image>"}, {"from": "gpt", "value": "women's clothes"}]
|
00000002395.jpg
|
2395_evjvqa
|
[{"from": "human", "value": "what are people sitting on the ground doing?\n<image>"}, {"from": "gpt", "value": "they are playing traditional musical instruments"}]
|
00000002395.jpg
|
2395_evjvqa
|
[{"from": "human", "value": "what are those people sitting on to play traditional musical instrument?\n<image>"}, {"from": "gpt", "value": "they are sitting on a mat"}]
|
00000002395.jpg
|
2395_evjvqa
|
[{"from": "human", "value": "what is the man wearing green shirt doing?\n<image>"}, {"from": "gpt", "value": "he is standing"}]
|
00000002396.jpg
|
2396_evjvqa
|
[{"from": "human", "value": "what are the costumes of thoses three people?\n<image>"}, {"from": "gpt", "value": "they are wearing red shirt with yellow star in the middle"}]
|
00000002396.jpg
|
2396_evjvqa
|
[{"from": "human", "value": "who is keeping the camera?\n<image>"}, {"from": "gpt", "value": "the camera is holding by the wan wearing hat"}]
|
00000004403.jpg
|
4403_evjvqa
|
[{"from": "human", "value": "where are these two boys sitting?\n<image>"}, {"from": "gpt", "value": "on a bench in a public place"}]
|
00000004403.jpg
|
4403_evjvqa
|
[{"from": "human", "value": "what are these two boys doing?\n<image>"}, {"from": "gpt", "value": "they are using phones"}]
|
00000004406.jpg
|
4406_evjvqa
|
[{"from": "human", "value": "what is the woman in the pink coat doing?\n<image>"}, {"from": "gpt", "value": "walking the motorbike"}]
|
00000004406.jpg
|
4406_evjvqa
|
[{"from": "human", "value": "what is the woman in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "she sales meat for the customer"}]
|
00000004406.jpg
|
4406_evjvqa
|
[{"from": "human", "value": "where is the woman in black overcoat standing?\n<image>"}, {"from": "gpt", "value": "at the meat stall in the supermarket"}]
|
00000004406.jpg
|
4406_evjvqa
|
[{"from": "human", "value": "what is the woman in the black coat going to buy?\n<image>"}, {"from": "gpt", "value": "the raw meat"}]
|
00000004406.jpg
|
4406_evjvqa
|
[{"from": "human", "value": "what do they sell at this stall?\n<image>"}, {"from": "gpt", "value": "the raw meat"}]
|
00000004407.jpg
|
4407_evjvqa
|
[{"from": "human", "value": "what does this man use the boat to carry?\n<image>"}, {"from": "gpt", "value": "the coconuts"}]
|
00000004407.jpg
|
4407_evjvqa
|
[{"from": "human", "value": "what means does this man carry coconuts?\n<image>"}, {"from": "gpt", "value": "the boat"}]
|
00000004407.jpg
|
4407_evjvqa
|
[{"from": "human", "value": "how are these coconuts transported?\n<image>"}, {"from": "gpt", "value": "by river"}]
|
00000004410.jpg
|
4410_evjvqa
|
[{"from": "human", "value": "what does the woman in the red coat do?\n<image>"}, {"from": "gpt", "value": "she sales the flowers"}]
|
00000004410.jpg
|
4410_evjvqa
|
[{"from": "human", "value": "what is the woman in the red coat doing?\n<image>"}, {"from": "gpt", "value": "she is giving the flowers to customers"}]
|
00000004410.jpg
|
4410_evjvqa
|
[{"from": "human", "value": "where is the bouquet stall openned by the woman?\n<image>"}, {"from": "gpt", "value": "on the side of the road"}]
|
End of preview. Expand
in Data Studio
README.md exists but content is empty.
- Downloads last month
- 8